Findings of the Association for Computational Linguistics: EMNLP 2021 2021
DOI: 10.18653/v1/2021.findings-emnlp.44
|View full text |Cite
|
Sign up to set email alerts
|

Semi-supervised Relation Extraction via Incremental Meta Self-Training

Abstract: To alleviate human efforts from obtaining large-scale annotations, Semi-Supervised Relation Extraction methods aim to leverage unlabeled data in addition to learning from limited samples. Existing self-training methods suffer from the gradual drift problem, where noisy pseudo labels on unlabeled data are in-

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
32
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 40 publications
(37 citation statements)
references
References 22 publications
0
32
0
Order By: Relevance
“…However, these works heavily rely on a frequently re-initialized linear classification layer which interferes with representation learning. Zhan et al (2020) proposes Online Deep Clustering that performs clustering and network update simultaneously rather than alternatingly to tackle this concern, however, the noisy pseudo labels still affect feature clustering when updating the network (Hu et al, 2021a;Li et al, 2022b;Lin et al, 2022).…”
Section: Related Workmentioning
confidence: 99%
“…However, these works heavily rely on a frequently re-initialized linear classification layer which interferes with representation learning. Zhan et al (2020) proposes Online Deep Clustering that performs clustering and network update simultaneously rather than alternatingly to tackle this concern, however, the noisy pseudo labels still affect feature clustering when updating the network (Hu et al, 2021a;Li et al, 2022b;Lin et al, 2022).…”
Section: Related Workmentioning
confidence: 99%
“…Conventional RE methods include supervised (Zelenko et al, 2002;Liu et al, 2013;Zeng et al, 2014;Miwa and Bansal, 2016), semi-supervised (Chen et al, 2006;Sun et al, 2011;Hu et al, 2020) and distantly supervised methods (Mintz et al, 2009;Yao et al, 2011;Zeng et al, 2015;Han et al, 2018a). These methods rely on a predefined relation set and have limitations in real scenario where novel relations are emerging.…”
Section: Related Workmentioning
confidence: 99%
“…Self-training has been studied for many years (Yarowsky, 1995;Riloff and Wiebe, 2003;Rosenberg et al, 2005) and widely adopted in many NLP tasks including speech recognition (Kahn et al, 2020;Park et al, 2020), parsing (McClosky et al, 2006McClosky and Charniak, 2008), and pre-training (Du et al, 2021). Self-Training suffers from inaccurate pseudo labels (Arazo et al, 2020(Arazo et al, , 2019Hu et al, 2021a) especially when the teacher model is trained on insufficient and unbalanced datasets. To address this problem, (Pham et al, 2020;Wang et al, 2021b;Hu et al, 2021a) propose to utilize the performance of the student model on the held out labeled data as a Meta Learning objective to update the teacher model or improve the pseudo-label generation process.…”
Section: Related Workmentioning
confidence: 99%
“…Self-Training suffers from inaccurate pseudo labels (Arazo et al, 2020(Arazo et al, , 2019Hu et al, 2021a) especially when the teacher model is trained on insufficient and unbalanced datasets. To address this problem, (Pham et al, 2020;Wang et al, 2021b;Hu et al, 2021a) propose to utilize the performance of the student model on the held out labeled data as a Meta Learning objective to update the teacher model or improve the pseudo-label generation process. (Hu et al, 2021b) leverage the cosine distance between gradients computed on labeled data and pseudolabeled data as feedback to guide the self-training process.…”
Section: Related Workmentioning
confidence: 99%