2022
DOI: 10.48550/arxiv.2201.12431
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval

Abstract: Retrieval-based language models (R-LM) model the probability of natural language text by combining a standard language model (LM) with examples retrieved from an external datastore at test time. While effective, a major bottleneck of using these models in practice is the computationally costly datastore search, which can be performed as frequently as every time step. In this paper, we present RETOMATON -retrieval automatonwhich approximates the datastore search, based on (1) clustering of entries into "states"… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 20 publications
0
3
0
Order By: Relevance
“…Recently, Retrieval-based language models (R-LMs) [1] have risen to improve over language models in various tasks such as unconditional language modeling [14,25], machine translation [10,24], text classification [27] and question answering [23]. R-LMs is a non-parametric method that uses training examples to augment the language model predictions at test time.…”
Section: Retrieval-based Language Modelsmentioning
confidence: 99%
“…Recently, Retrieval-based language models (R-LMs) [1] have risen to improve over language models in various tasks such as unconditional language modeling [14,25], machine translation [10,24], text classification [27] and question answering [23]. R-LMs is a non-parametric method that uses training examples to augment the language model predictions at test time.…”
Section: Retrieval-based Language Modelsmentioning
confidence: 99%
“…Recently, retrieval-augmented LMs have shown a series of impressive results (Grave et al, 2017;Guu et al, 2018;He et al, 2020;Khandelwal et al, 2020b;Borgeaud et al, 2022;Alon et al, 2022). Retrieval-augmented LMs compute next token distributions based not only on the immediately preceding context c t and the model parameters, but also on an external datastore, from which examples are retrieved and incorporated into the base LM's prediction.…”
Section: Introductionmentioning
confidence: 99%
“…Finally, one significant drawback to the current kNN-LM is the inefficiency of kNN search performed at each step (He et al, 2021;Borgeaud et al, 2022;Alon et al, 2022;Wang et al, 2022). Because of the similarity between kNN-LM and the parametric LM's last layers and the many design choices, we also demonstrate that we are able to make kNN-LM more efficient by substituting the kNN search with another matrix operation that can fit in accelerator memory while maintaining more than half the perplexity improvement, or more than 6.5% relative improvement compared to the base LM.…”
Section: Introductionmentioning
confidence: 99%