Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 2021
DOI: 10.18653/v1/2021.findings-acl.263
|View full text |Cite
|
Sign up to set email alerts
|

Inspecting the concept knowledge graph encoded by modern language models

Abstract: The field of natural language understanding has experienced exponential progress in the last few years, with impressive results in several tasks. This success has motivated researchers to study the underlying knowledge encoded by these models. Despite this, attempts to understand their semantic capabilities have not been successful, often leading to non-conclusive, or contradictory conclusions among different works. Via a probing classifier, we extract the underlying knowledge graph of nine of the most influen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 45 publications
0
1
0
Order By: Relevance
“…However, it is not clear in which layers it is crucial to have the PC mechanism. We hypothesize that this is related to the fact that the BERT-style models encode syntactic and semantic features in different layers (Jawahar et al, 2019;Aspillaga et al, 2021), so a specialized PC mechanism for syntax or semantics would be desirable. We left this study for future work.…”
Section: Ablation Studymentioning
confidence: 99%
“…However, it is not clear in which layers it is crucial to have the PC mechanism. We hypothesize that this is related to the fact that the BERT-style models encode syntactic and semantic features in different layers (Jawahar et al, 2019;Aspillaga et al, 2021), so a specialized PC mechanism for syntax or semantics would be desirable. We left this study for future work.…”
Section: Ablation Studymentioning
confidence: 99%
“…In particular, they found that neurons in XLNet are more localized in encoding individual linguistic information compared to BERT, where neurons are shared across multiple properties. By adopting the method of Hewitt and Manning (2019), Aspillaga et al (2021) investigated whether pre-trained language models encode semantic information, for instance by checking their representations against the lexico-semantic structure of WordNet (Miller, 1994).…”
Section: Related Workmentioning
confidence: 99%