Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.143
|View full text |Cite
|
Sign up to set email alerts
|

Contextual Rephrase Detection for Reducing Friction in Dialogue Systems

Abstract: For voice assistants like Alexa, Google Assistant and Siri, correctly interpreting users' intentions is of utmost importance. However, users sometimes experience friction with these assistants, caused by errors from different system components or user errors such as slips of the tongue. Users tend to rephrase their query until they get a satisfactory response. Rephrase detection is used to identify the rephrases and has long been treated as a task with pairwise input, which does not fully utilize the contextua… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 10 publications
0
3
0
Order By: Relevance
“…For re-ranking, given a pair of utterance and entity, we concatenate the output vector of CLS token of RoBERTa and the pooling output vector of GAT, and pass them to an MLP layer to produce the relevance score of the pair. For corrupt entity span detection, we predict the span's start and end positions at the token level, following similar approaches such as in [16] and [17]. Specifically, assume W S and W E are the start and the end vector respectively, and T i ∈ R H is the final hidden vector for the i th input token, then the score of a candidate span from position i to position j is computed as:…”
Section: L2 Re-ranking + Span Detectionmentioning
confidence: 99%
“…For re-ranking, given a pair of utterance and entity, we concatenate the output vector of CLS token of RoBERTa and the pooling output vector of GAT, and pass them to an MLP layer to produce the relevance score of the pair. For corrupt entity span detection, we predict the span's start and end positions at the token level, following similar approaches such as in [16] and [17]. Specifically, assume W S and W E are the start and the end vector respectively, and T i ∈ R H is the final hidden vector for the i th input token, then the score of a candidate span from position i to position j is computed as:…”
Section: L2 Re-ranking + Span Detectionmentioning
confidence: 99%
“…For textual data generation, Rik et al [23] used a Transformer to generate text from a knowledge graph. Zhuoyi et al [24] designed a novel rephrase detection system based on contextual content in dialogue scenes. Image generation and synthesis techniques are widely utilized.…”
Section: Attack and Defense In Other Layersmentioning
confidence: 99%
“…The information in these branches is fused to obtain a richer representation. Many other tasks [53] also use an attention mechanism to perform context modeling for better representations. For example, ContextNet [54] and Dual-mode ASR [55] that propose a novel CNN-RNN-transducer architecture with global context information for speech recognition, Cp-GAN for speech enhancement [56], and context aware attention in speech emotion detection [57].…”
mentioning
confidence: 99%