2020
DOI: 10.1609/aaai.v34i05.6407
|View full text |Cite
|
Sign up to set email alerts
|

Are Noisy Sentences Useless for Distant Supervised Relation Extraction?

Abstract: The noisy labeling problem has been one of the major obstacles for distant supervised relation extraction. Existing approaches usually consider that the noisy sentences are useless and will harm the model's performance. Therefore, they mainly alleviate this problem by reducing the influence of noisy sentences, such as applying bag-level selective attention or removing noisy sentences from sentence-bags. However, the underlying cause of the noisy labeling problem is not the lack of useful information, but the m… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
18
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 33 publications
(22 citation statements)
references
References 12 publications
1
18
0
Order By: Relevance
“…In MIL, the training and testing processes are performed at the bag level, where a bag contains noisy sentences mentioning the same entity pair but possibly not describing the same relation. Studies using MIL can be broadly classified into two categories: 1) the soft de-noise methods that leverage soft weights to differentiate the influence of each sentence (Lin et al, 2016;Han et al, 2018c;Li et al, 2020;Hu et al, 2019a;Ye and Ling, 2019;Yuan et al, 2019a,b); 2) the hard de-noise methods that remove noisy sentences from the bag (Zeng et al, 2015;Qin et al, 2018;Han et al, 2018a;Shang, 2019). However, these bag-level approaches fail to map each sentence inside bags with explicit sentence labels.…”
Section: Bag Labels (We Need) Sentence Labelsmentioning
confidence: 99%
“…In MIL, the training and testing processes are performed at the bag level, where a bag contains noisy sentences mentioning the same entity pair but possibly not describing the same relation. Studies using MIL can be broadly classified into two categories: 1) the soft de-noise methods that leverage soft weights to differentiate the influence of each sentence (Lin et al, 2016;Han et al, 2018c;Li et al, 2020;Hu et al, 2019a;Ye and Ling, 2019;Yuan et al, 2019a,b); 2) the hard de-noise methods that remove noisy sentences from the bag (Zeng et al, 2015;Qin et al, 2018;Han et al, 2018a;Shang, 2019). However, these bag-level approaches fail to map each sentence inside bags with explicit sentence labels.…”
Section: Bag Labels (We Need) Sentence Labelsmentioning
confidence: 99%
“…Apart from that, methods treat the identified potentially noisy samples differently. They are either kept for further training with reduced weights (Jat et al, 2018;He et al, 2020), corrected (Shang, 2019) or eliminated (Qin et al, 2018). Thus, denoising methods vary significantly depending on the data and task, what makes the creation of a platform for comparison crucial.…”
Section: Weak Supervisionmentioning
confidence: 99%
“…In the field of distantly supervised relation extraction, the multi-instance learning framework is introduced to handle the label noise of DS. Recently, MIL has become a common paradigm for DSRE and many efforts have been made for further improvements (Lin et al, 2016;Qin et al, 2018;Ye and Ling, 2019;Huang and Du, 2019;Shang et al, 2020). In these MIL frameworks, sentences are first encoded by handcrafted features (Mintz et al, 2009;Hoffmann et al, 2011) or neural networks.…”
Section: Related Workmentioning
confidence: 99%
“…We evaluate our model on the widely used DSRE dataset -NYT (Mintz et al, 2009), which aligns Freebase (Bollacker et al, 2008) (Huang and Du, 2019;Shang et al, 2020) since then adopt the correct dataset as the benchmark. In order to ensure the fairness and scientificity of our experiments, we use the original dataset release and employ the popular RE toolkit OpenNRE (Han et al, 2019) in our study .…”
Section: Datasetmentioning
confidence: 99%
See 1 more Smart Citation