2022
DOI: 10.48550/arxiv.2210.01963
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

COMPS: Conceptual Minimal Pair Sentences for testing Robust Property Knowledge and its Inheritance in Pre-trained Language Models

Abstract: A characteristic feature of human semantic memory is its ability to not only store and retrieve the properties of concepts observed through experience, but to also facilitate the inheritance of properties (can breathe) from superordinate concepts (ANIMAL) to their subordinates (DOG)-i.e. demonstrate property inheritance. In this paper, we present COMPS, a collection of minimal pair sentences that jointly tests pre-trained language models (PLMs) on their ability to attribute properties to concepts and their abi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 37 publications
0
2
0
Order By: Relevance
“…For instance, when prompted with "Seinfeld premiered on ____" and "Seinfeld originally aired on ____", they might provide names of different TV networks despite the fact that the prompts have the same semantic content Elazar et al [see also Ravichander et al 2020Ravichander et al 2020. Misra et al [2022] show another interesting failure of even advanced models like GPT-3 on a task where they are asked to reason about properties of novel objects. Models generally learn properties of objects, like that robins can fly and penguins cannot.…”
Section: World Knowledge and Commonsense Reasoningmentioning
confidence: 99%
See 1 more Smart Citation
“…For instance, when prompted with "Seinfeld premiered on ____" and "Seinfeld originally aired on ____", they might provide names of different TV networks despite the fact that the prompts have the same semantic content Elazar et al [see also Ravichander et al 2020Ravichander et al 2020. Misra et al [2022] show another interesting failure of even advanced models like GPT-3 on a task where they are asked to reason about properties of novel objects. Models generally learn properties of objects, like that robins can fly and penguins cannot.…”
Section: World Knowledge and Commonsense Reasoningmentioning
confidence: 99%
“…That is, given the example in the preceding sentence as input, the model should be able to know that "I went to a blick" is more probable than "I went to a dax" since blick was used as a noun. They conclude that BERT succeeds partially at this task: it does learn to generalize, but only after repeated examples [but see , Misra et al, 2022 for ways in which the word itself affects compositional ability]. More recent models, such as GPT-3, seem to be able to use a novel word appropriately right away, at least if prompted correctly , McCoy et al, 2021b.…”
Section: Llms Learn Abstractionsmentioning
confidence: 99%