Proceedings of the 24th Conference on Computational Natural Language Learning 2020
DOI: 10.18653/v1/2020.conll-1.45
|View full text |Cite
|
Sign up to set email alerts
|

Are Pretrained Language Models Symbolic Reasoners over Knowledge?

Abstract: How can pretrained language models (PLMs) learn factual knowledge from the training set? We investigate the two most important mechanisms: reasoning and memorization. Prior work has attempted to quantify the number of facts PLMs learn, but we present, using synthetic data, the first study that investigates the causal relation between facts present in training and facts learned by the PLM. For reasoning, we show that PLMs seem to learn to apply some symbolic reasoning rules correctly but struggle with others, i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
22
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 22 publications
(24 citation statements)
references
References 24 publications
2
22
0
Order By: Relevance
“…Petroni et al (2019) demonstrate that language models are able to recall factual knowledge without any fine-tuning and can somewhat function as an unsupervised ODQA system. However, our experiments suggest that, large-scale language models (when fine-tuned to directly answer questions using a set of training QA pairs) struggle to answer questions about low frequency entities and relations, similar to the findings of Kassner et al (2020) and Dufter et al (2021).…”
Section: Model Category Analysissupporting
confidence: 83%
“…Petroni et al (2019) demonstrate that language models are able to recall factual knowledge without any fine-tuning and can somewhat function as an unsupervised ODQA system. However, our experiments suggest that, large-scale language models (when fine-tuned to directly answer questions using a set of training QA pairs) struggle to answer questions about low frequency entities and relations, similar to the findings of Kassner et al (2020) and Dufter et al (2021).…”
Section: Model Category Analysissupporting
confidence: 83%
“…Recently, pre-trained language models (Peters et al, 2018;Devlin et al, 2019;Brown et al, 2020) have achieved promising performance on many NLP tasks. Apart from utilizing the universal representations from pre-trained models in downstream tasks, some literatures have shown the potential of pretrained masked language models (e.g., BERT (Devlin et al, 2019) and RoBERTa (Liu et al, 2019b)) to be factual knowledge bases (Petroni et al, 2019;Bouraoui et al, 2020;Jiang et al, 2020b;Shin et al, 2020;Jiang et al, 2020a;Wang et al, 2020;Kassner and Schütze, 2020a;Kassner et al, 2020). For example, to extract the birthplace of Steve Jobs, we can query MLMs like BERT with "Steve Jobs was born in [MASK]", where Steve Jobs is the subject of the fact, "was born in" is a prompt string for the relation "place-of-birth" and [MASK] is a placeholder for the object to predict.…”
Section: Introductionmentioning
confidence: 99%
“…generalising an observation about a given concept to hyponyms of that concept. In [58], the ability of transformer based LMs to generalise observed facts is analysed in a systematic way, by training an LM from scratch on a synthetic corpus in which various regularities are present. They find that LMs are indeed capable of discovering symbolic rules, and capable of applying such rules for inferring facts not present in the training corpus, although they also identified important limitations.…”
Section: Contextualised Language Models As Rule-based Reasonersmentioning
confidence: 99%