2022
DOI: 10.48550/arxiv.2202.07646
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Quantifying Memorization Across Neural Language Models

Abstract: Large language models (LMs) have been shown to memorize parts of their training data, and when prompted appropriately, they will emit the memorized training data verbatim. This is undesirable because memorization violates privacy (exposing user data), degrades utility (repeated easy-to-memorize text is often low quality), and hurts fairness (some texts are memorized over others).We describe three log-linear relationships that quantify the degree to which LMs emit memorized training data. Memorization significa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

5
73
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
4

Relationship

1
7

Authors

Journals

citations
Cited by 47 publications
(102 citation statements)
references
References 19 publications
(39 reference statements)
5
73
0
Order By: Relevance
“…We find that the attack's performance increases steadily with the number of tokens known to the adversary. This mirrors the findings in [12], who show that prompting a language model with longer prefixes increases the likelihood of extracting memorized content. As long as the attacker knows more than 𝑛 = 8 tokens of context (6 English words on average), they increase exposure of secrets by poisoning the model.…”
Section: Attacks With Relaxed Capabilitiessupporting
confidence: 83%
“…We find that the attack's performance increases steadily with the number of tokens known to the adversary. This mirrors the findings in [12], who show that prompting a language model with longer prefixes increases the likelihood of extracting memorized content. As long as the attacker knows more than 𝑛 = 8 tokens of context (6 English words on average), they increase exposure of secrets by poisoning the model.…”
Section: Attacks With Relaxed Capabilitiessupporting
confidence: 83%
“…Do memory architectures improve performance in rare situations? While transformer architectures have enabled large language models to improve performance on rare experiences (Carlini et al, 2022), we saw little evidence that changing the agent's core memory from a LSTM to a transformer led to better performance on rare items. There are many differences in the experience, objective, and scale that could potentially explain this difference-for example modern language models can condition on words across many consecutive sentences, while the IMPALA and V-MPO algorithms do not enable an agent to condition on stimuli outside of the current episode.…”
Section: Methodsmentioning
confidence: 85%
“…In particular, transformers (Vaswani et al, 2017) have shown substantial ability to learn about rare events, with the largest language models exhibiting recall of some rare training experiences (e.g. Carlini et al, 2022). While evaluating such a large architecture would be prohibitive, we compare to agents with a Gated TransformerXL memory (Parisotto et al, 2020), to evaluate whether the memory architecture can affect learning from rare experiences.…”
Section: Reinforcement Learning From Rare Experiencesmentioning
confidence: 99%
“…This finding is consistent with several concurrent works, which show similar connections in GPT-based models. These works study the impact of duplication of training sequence on regeneration of the sequence (Carlini et al, 2022;Kandpal et al, 2022), and the effect on few-shot numerical reasoning (Razeghi et al, 2022). One explanation for this phenomenon is the increase in the expected number of times labels are masked during pretraining.…”
Section: Which Factors Affect Exploitation?mentioning
confidence: 99%
“…These works mostly use GPT-based models. Carlini et al (2022) showed that memorization of language models grows with model size, training data duplicates, and the prompt length. They further found that masked language models memorize an order of magnitude less data compared to causal language model.…”
Section: Related Workmentioning
confidence: 99%