Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2018
DOI: 10.18653/v1/p18-1198
|View full text |Cite
|
Sign up to set email alerts
|

What you can cram into a single $&!#* vector: Probing sentence embeddings for linguistic properties

Abstract: Although much effort has recently been devoted to training high-quality sentence embeddings, we still have a poor understanding of what they are capturing. "Downstream" tasks, often based on sentence classification, are commonly used to evaluate the quality of sentence representations. The complexity of the tasks makes it however difficult to infer what kind of information is present in the representations. We introduce here 10 probing tasks designed to capture simple linguistic features of sentences, and we u… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

9
611
0
2

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 598 publications
(683 citation statements)
references
References 26 publications
9
611
0
2
Order By: Relevance
“…Through probing methods, it has been shown that a broad range of supervised learning tasks can be turned into tools for understanding the properties of contextual word representations (Conneau et al, 2018;. Alain and Bengio (2016) suggested we may think of probes as "thermometers used to measure the temperature simultaneously at many different locations".…”
Section: Resultsmentioning
confidence: 99%
“…Through probing methods, it has been shown that a broad range of supervised learning tasks can be turned into tools for understanding the properties of contextual word representations (Conneau et al, 2018;. Alain and Bengio (2016) suggested we may think of probes as "thermometers used to measure the temperature simultaneously at many different locations".…”
Section: Resultsmentioning
confidence: 99%
“…Another branch of work uses probing tasks in which the objective is to predict the value of a particular linguistic feature given an input sentence. Probing tasks have been used to investigate whether sentence embeddings encode syntactic and surface features such as tense and voice (Shi et al, 2016), sentence length and word content (Adi et al, 2016), or syntactic depth and morphological number (Conneau et al, 2018). Giulianelli et al (2018) use diagnostic classifiers to track the propagation of information in RNNbased language models.…”
Section: Introductionmentioning
confidence: 99%
“…More recently, it was shown that both i-vectors and x-vectors contain information about the speaking style and emotion [12]. In natural language processing (NLP), probing tasks for embeddings have gained attention [13] due to sentence encoders such as BERT [14], which are pretrained on language modeling, but achieve state-of-the-art performance across several other tasks.…”
Section: Introductionmentioning
confidence: 99%