2021
DOI: 10.1162/tacl_a_00356
|View full text |Cite
|
Sign up to set email alerts
|

Augmenting Transformers with KNN-Based Composite Memory for Dialog

Abstract: Various machine learning tasks can benefit from access to external information of different modalities, such as text and images. Recent work has focused on learning architectures with large memories capable of storing this knowledge. We propose augmenting generative Transformer neural networks with KNN-based Information Fetching (KIF) modules. Each KIF module learns a read operation to access fixed external knowledge. We apply these modules to generative dialog modeling, a challenging task where information mu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
19
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 26 publications
(28 citation statements)
references
References 21 publications
(39 reference statements)
0
19
0
Order By: Relevance
“…Retrieval-augmented models. Retrievalaugmented models are now widely adopted in open-domain question answering (Chen et al, 2017;de Masson d'Autume et al, 2019;Izacard and Grave, 2021), dialogue (Dinan et al, 2019;Fan et al, 2021;Thulke et al, 2021) and machine translation (Bapna and Firat, 2019;Khandelwal et al, 2020a). We focus on retrieval augmentation for language modelling (Merity et al, 2017;Grave et al, 2016;Khandelwal et al, 2020b;Yogatama et al, 2021).…”
Section: Knowledge-enhancedmentioning
confidence: 99%
“…Retrieval-augmented models. Retrievalaugmented models are now widely adopted in open-domain question answering (Chen et al, 2017;de Masson d'Autume et al, 2019;Izacard and Grave, 2021), dialogue (Dinan et al, 2019;Fan et al, 2021;Thulke et al, 2021) and machine translation (Bapna and Firat, 2019;Khandelwal et al, 2020a). We focus on retrieval augmentation for language modelling (Merity et al, 2017;Grave et al, 2016;Khandelwal et al, 2020b;Yogatama et al, 2021).…”
Section: Knowledge-enhancedmentioning
confidence: 99%
“…Recent efforts in NLP have shown the effectiveness of relying on an explicit set of nearest neighbors to be effective for language modelling , question answering (Kassner and Schütze, 2020) and knowledge-grounded dialog (Fan et al, 2020). However, these approaches condition on examples only during inference or in a non end-to-end manner.…”
Section: Example-driven Trainingmentioning
confidence: 99%
“…Marino et al (2019) introduced OK-VQA, a novel VQA dataset that requires the use of an external KS. Fan et al (2020) applied a KS to multi-modal dialogue. In our work, we focus on a more naturally aligned KS, in the form of images and captions, which better reflects the data generated in newspapers and social media.…”
Section: Related Workmentioning
confidence: 99%