2021 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU) 2021
DOI: 10.1109/asru51503.2021.9688296
|View full text |Cite
|
Sign up to set email alerts
|

Audio Embeddings Help to Learn Better Dialogue Policies

Abstract: Neural transformer architectures have gained a lot of interest for text-based dialogue management in the last few years. They have shown high learning capabilities for open domain dialogue with huge amounts of data and also for domain adaptation in task-oriented setups. But the potential benefits of exploiting the users' audio signal have rarely been explored in such frameworks. In this work, we combine text dialogue history representations generated by a GPT-2 model with audio embeddings obtained by the recen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(5 citation statements)
references
References 18 publications
0
5
0
Order By: Relevance
“…Last, this study presents several novelties compared to our preliminary work [5]. We provide a larger experimentation and a much deeper analysis of how, when and why speech representations help to learn better dialogue policies.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…Last, this study presents several novelties compared to our preliminary work [5]. We provide a larger experimentation and a much deeper analysis of how, when and why speech representations help to learn better dialogue policies.…”
Section: Related Workmentioning
confidence: 99%
“…User Audio Sampler. Since audio signals need to be fed to the proposed dialogue policies, we employ the User Audio Sampler proposed in [5] to sample an audio turn from the corpus. First, it selects output candidates filtering the audios of dialogue turns labelled with the same dialogue acts and slots generated by the UM.…”
Section: B Dialogue Pipeline For Simulationsmentioning
confidence: 99%
See 2 more Smart Citations
“…DRL agents are often trained from scratch instead of inheriting useful behaviours from other agents. Some agents from Table 4 [such as (Williams and Zweig 2016;Liu et al 2017;Zorrilla et al 2021)] have avoided learning from scratch by showing that applying DRL on top of non-DRL or supervised methods yields improved performance due to the optimisation element that DRL brings instead of only mimicking demonstration data. But those systems typically focus a single dataset and the idea of transferring useful and effective knowledge from other/many tasks to a new or targeted task remains to be demonstrated.…”
Section: Knowledge Transfer and Generalisationmentioning
confidence: 99%