Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing 2016
DOI: 10.18653/v1/d16-1235
|View full text |Cite
|
Sign up to set email alerts
|

SimVerb-3500: A Large-Scale Evaluation Set of Verb Similarity

Abstract: Verbs play a critical role in the meaning of sentences, but these ubiquitous words have received little attention in recent distributional semantics research. We introduce SimVerb-3500, an evaluation resource that provides human ratings for the similarity of 3,500 verb pairs. SimVerb-3500 covers all normed verb types from the USF free-association database, providing at least three examples for every Verb-Net class. This broad coverage facilitates detailed analyses of how syntactic and semantic phenomena togeth… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

9
202
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 183 publications
(214 citation statements)
references
References 31 publications
9
202
0
Order By: Relevance
“…First we consider a group of word-level similarity datasets that are commonly used as benchmarks in previous research: WS-353-SIM (Finkelstein et al, 2001), YP-130 (Yang and Powers, 2005), SIMLEX-999 (Hill et al, 2015), SimVerb-3500 (Gerz et al, 2016), RW-STANFORD (Luong Table 1: Spearman's ρ on word similarity tasks for combinations of word vectors and the following similarity metrics: cosine similarity (COS), Pearson's r (PRS), Spearman's ρ (SPR), and Kendall τ (KEN). N indicates the proportion of sentence vectors in a task for which the null hypothesis of normality in a Shapiro-Wilk test was not rejected at α = 0.05.…”
Section: Methodsmentioning
confidence: 99%
“…First we consider a group of word-level similarity datasets that are commonly used as benchmarks in previous research: WS-353-SIM (Finkelstein et al, 2001), YP-130 (Yang and Powers, 2005), SIMLEX-999 (Hill et al, 2015), SimVerb-3500 (Gerz et al, 2016), RW-STANFORD (Luong Table 1: Spearman's ρ on word similarity tasks for combinations of word vectors and the following similarity metrics: cosine similarity (COS), Pearson's r (PRS), Spearman's ρ (SPR), and Kendall τ (KEN). N indicates the proportion of sentence vectors in a task for which the null hypothesis of normality in a Shapiro-Wilk test was not rejected at α = 0.05.…”
Section: Methodsmentioning
confidence: 99%
“…A score close to 1 indicates an embedding close to the human judgement. We use MC-30 (Miller and Charles, 1991), MEN (Bruni et al, 2014), MTurk-287 (Radinsky et al, 2011), MTurk-771 (Halawi et al, 2012), RG-65 (Rubenstein and Goodenough, 1965), RW (Luong et al, 2013), SimVerb-3500 (Gerz et al, 2016), WordSim-353 (Finkelstein et al, 2001) and YP-130 (Yang and Powers, 2006) classic datasets. We follow the same protocol used by Word2vec and fastText by discarding pairs which contain a word that is not in our embedding.…”
Section: Word Similarity Evaluationmentioning
confidence: 99%
“…(Elman, 2004). The new models go much further by capturing a considerable amount of variance of human word-to-word similarity ratings (e.g., Gerz, Vulić, Hill, Reichart, & Korhonen, 2016;Levy & Goldberg, 2014). Here are some similarity relations word2vec captures by simply attempting to predict words from surrounding words:…”
mentioning
confidence: 99%