2011
DOI: 10.5715/jnlp.18.153
|View full text |Cite
|
Sign up to set email alerts
|

Active Learning with Subsequence Sampling Strategy for Sequence Labeling Tasks

Abstract: We propose an active learning framework for sequence labeling tasks. In each iteration, a set of subsequences are selected and manually labeled, while the other parts of sequences are left unannotated. The learning will stop automatically when the training data between consecutive iterations does not significantly change. We evaluate the proposed framework on chunking and named entity recognition data provided by CoNLL. Experimental results show that we succeed in obtaining the supervised F 1 only with 6.98%, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2014
2014
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(14 citation statements)
references
References 12 publications
(16 reference statements)
0
14
0
Order By: Relevance
“…This way, the annotators only need to assign types to the chosen subspans without having to read and annotate the full sequence. To cope with the resulting partial annotation of sequences, we apply a constrained version of conditional random fields (CRFs), partial CRFs, during training that only learn from the annotated subspans (Tsuboi et al, 2008;Wanvarie et al, 2011). To evaluate our proposed methods, we conducted simulated active learning experiments on 5 languages: Spanish, Dutch, German, Hindi and Indonesian.…”
Section: Introductionmentioning
confidence: 99%
“…This way, the annotators only need to assign types to the chosen subspans without having to read and annotate the full sequence. To cope with the resulting partial annotation of sequences, we apply a constrained version of conditional random fields (CRFs), partial CRFs, during training that only learn from the annotated subspans (Tsuboi et al, 2008;Wanvarie et al, 2011). To evaluate our proposed methods, we conducted simulated active learning experiments on 5 languages: Spanish, Dutch, German, Hindi and Indonesian.…”
Section: Introductionmentioning
confidence: 99%
“…The main AL works in this latter line of work are (Shen et al, 2004), ) and (Wanvarie et al, 2011). Shen et al (2004) adopted SVMs as learning algorithm and proposed two strategies that combine three criteria, informativeness, representativeness and diversity.…”
Section: Related Workmentioning
confidence: 99%
“…At this point an AL strategy S will select a number of examples B that once labeled will hopefully improve the performance of the next classifier Φ i+1 . Algorithm 1 shows the pool-based AL framework for partially annotated sequences as introduced in (Wanvarie et al, 2011). Differently from AL for fully labeled sequences , thanks to the finer granularity of the partially labeled model, we use the token as basic annotation unit, instead of the entire sequence.…”
Section: Active Learning Strategiesmentioning
confidence: 99%
See 2 more Smart Citations