Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.109
|View full text |Cite
|
Sign up to set email alerts
|

How Relevant Are Selectional Preferences for Transformer-based Language Models?

Abstract: Selectional preference is defined as the tendency of a predicate to favour particular arguments within a certain linguistic context, and likewise, reject others that result in conflicting or implausible meanings. The stellar success of contextual word embedding models such as BERT in NLP tasks has led many to question whether these models have learned linguistic information, but up till now, most research has focused on syntactic information. We investigate whether BERT contains information on the selectional … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(9 citation statements)
references
References 32 publications
0
5
0
Order By: Relevance
“…To verify and extend our findings, future work should test LLMs’ knowledge of selectional restrictions on features other than animacy, such as the physical constraints that a predicate places on its patients (Wang, Durrett, & Erk, 2018), evaluate their performance on impossible events that do not violate selectional restrictions per se (e.g., She gave birth to her mother, The man was killed twice , or After 10 coin tosses, she got 12 heads . ), and conduct more targeted tests of agent‐verb and patient‐verb plausibility (Metheniti et al., 2020).…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…To verify and extend our findings, future work should test LLMs’ knowledge of selectional restrictions on features other than animacy, such as the physical constraints that a predicate places on its patients (Wang, Durrett, & Erk, 2018), evaluate their performance on impossible events that do not violate selectional restrictions per se (e.g., She gave birth to her mother, The man was killed twice , or After 10 coin tosses, she got 12 heads . ), and conduct more targeted tests of agent‐verb and patient‐verb plausibility (Metheniti et al., 2020).…”
Section: Discussionmentioning
confidence: 99%
“…Matsuki et al., 2011). Computational evidence suggests that BERT models are able to generalize their knowledge of selectional restrictions in novel word‐learning paradigms (Thrush et al., 2020) and can partially rely on the semantics of the head predicate to predict upcoming event participants (Metheniti, Van de Cruys, & Hathout, 2020). The asymmetry in performance on possible/impossible versus likely/unlikely events was independent from the specifics of LLM architecture and training and was additionally present, in an even more marked way, in our baseline models.…”
Section: Discussionmentioning
confidence: 99%
“…Past work has probed token embeddings for knowledge of argument structure (Kann et al, 2019;Pavlick, 2022;Sasano and Korhonen, 2020;Tenney et al, 2019a,b;Zhu and de Melo, 2020). Other work has focused on neural networks' ability to predict the likelihood of a verb or noun in forms of an argument structure alternation (Chowdhury and Zamparelli, 2019;Hawkins et al, 2020b;Loáiciga et al, 2021;Metheniti et al, 2020;Petty et al, 2022;Yoshida and Oseki, 2022), and whether LLMs distinguish plausible from implausible argument-role mappings in role-reversal sentences (Ettinger, 2020). Though revealing of Type 0 knowledge, that work does not address whether LLMs can apply such knowledge productively, which is what drives our study.…”
Section: Related Workmentioning
confidence: 99%
“…It includes experiments concerning their correlation with human judgment in terms of semantic similarity or their ability to classify word pairs according to different types of relations (Chersoni et al, 2016;Xiang et al, 2020). Other studies also considered contextual embeddings from a semantic viewpoint but in more specific contexts: the impact of their training objectives (Mickus et al, 2020), their level of contextualization (Ethayarajh, 2019), their possible biases (Bommasani et al, 2020), their ability to represent word senses (Coenen et al, 2019), to build representations for rare words (Schick and Schütze, 2020), to account for selectional preferences (Metheniti et al, 2020) or to interpret logical metonymy (Rambelli et al, 2020). Finally, (Chronis and Erk, 2020) is linked to the representation of word senses through the notion of prototype but mainly applies it for characterizing semantic similarity versus semantic relatedness and abstractness versus concreteness.…”
Section: Semantic Study Of Contextual Word Embeddingsmentioning
confidence: 99%