2019
DOI: 10.48550/arxiv.1906.03158
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Matching the Blanks: Distributional Similarity for Relation Learning

Abstract: General purpose relation extractors, which can model arbitrary relations, are a core aspiration in information extraction. Efforts have been made to build general purpose extractors that represent relations with their surface forms, or which jointly embed surface forms with relations from an existing knowledge graph. However, both of these approaches are limited in their ability to generalize. In this paper, we build on extensions of Harris' distributional hypothesis to relations, as well as recent advances in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
20
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 24 publications
(33 citation statements)
references
References 10 publications
(15 reference statements)
0
20
0
Order By: Relevance
“…This vocabulary size is two orders of magnitude larger than in previous work that applies a Transformer model with full softmax loss [8,23,18]. Other works, such as [24] and [17], train a Transformer model with a large number of entities using sampled softmax, with either in-batch or in-example negative sampling. But as we shall show, sampled softmax, even with a large number of 128K negative samples, results in much worse quality.…”
Section: Wikipedia Entity Predictionmentioning
confidence: 99%
“…This vocabulary size is two orders of magnitude larger than in previous work that applies a Transformer model with full softmax loss [8,23,18]. Other works, such as [24] and [17], train a Transformer model with a large number of entities using sampled softmax, with either in-batch or in-example negative sampling. But as we shall show, sampled softmax, even with a large number of 128K negative samples, results in much worse quality.…”
Section: Wikipedia Entity Predictionmentioning
confidence: 99%
“…For an input sentence containing a relation mention, two entities Entity 1 and Entity 2 are marked in advance. We follow the labeling mechanism adopted by Soares et al (2019) and Zhang and Wang (2015) to enhance the position information of entities. For each sentence X = [x 1 , .., x T ], four reserved tokens…”
Section: Relation Classification Networkmentioning
confidence: 99%
“…FewRel 2.0 (Gao et al, 2019b) extends the dataset with few-shot domain adaption and few-shot none-of-the-above detection. Many works on FewRel datasets focus on improvement of methods, including modeling the distance distribution (Gao et al, 2019a;Ding et al, 2021), utilizing external knowledge such as knowledge graph (Qu et al, 2020), learning different level of features (Sun et al, 2019;Ye and Ling, 2019), and using pre-trained language models (Soares et al, 2019). Apart from method innovation on the standard setting of consistent few-shot RC, the investigation for inconsistent few-shot RC still in demand.…”
Section: Related Workmentioning
confidence: 99%