2018 International Conference on Asian Language Processing (IALP) 2018
DOI: 10.1109/ialp.2018.8629107
|View full text |Cite
|
Sign up to set email alerts
|

Learning How to Self-Learn: Enhancing Self-Training Using Neural Reinforcement Learning

Abstract: Self-training is a useful strategy for semisupervised learning, leveraging raw texts for enhancing model performances. Traditional self-training methods depend on heuristics such as model confidence for instance selection, the manual adjustment of which can be expensive. To address these challenges, we propose a deep reinforcement learning method to learn the self-training strategy automatically. Based on neural network representation of sentences, our model automatically learns an optimal policy for instance … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(6 citation statements)
references
References 33 publications
0
6
0
Order By: Relevance
“…A somewhat similar approach is self-learning -training on examples labelled by a model itself. While it is ineffective in many settings, [3] shows that it can improve results of few-shot NER task when combined with reinforcement learning.…”
Section: Related Workmentioning
confidence: 99%
“…A somewhat similar approach is self-learning -training on examples labelled by a model itself. While it is ineffective in many settings, [3] shows that it can improve results of few-shot NER task when combined with reinforcement learning.…”
Section: Related Workmentioning
confidence: 99%
“…Kumar et al (2010); Ma et al (2017); Li et al (2019); Mukherjee and Awadallah (2020) proposed to learn sampling weights for unlabeled data to control the selection process. Reinforcement learning (RL) methods (Chen et al, 2018;Wu et al, 2018; designed an additional Q-agent as the sample selector. Nevertheless, methods using learnable weights or RL provide marginal benefits compared to the elevated optimization cost.…”
Section: Self-trainingmentioning
confidence: 99%
“…Unlike classification tasks, noisy self-labeled data can be easily eliminated by removing those which have low confidence scores; there is a lack of a comprehensive means to determine this score for a sequence-labeling data point. In several recent re-search, a deep reinforcement learning (Chen et al, 2018) and meta-learning (Wang et al, 2020) has been proposed to reduce "error propagation from noisy pseudo-labels" for sequence labeling tasks.…”
Section: Self-trainingmentioning
confidence: 99%