2022
DOI: 10.1007/s11063-021-10737-x
|View full text |Cite
|
Sign up to set email alerts
|

LTP: A New Active Learning Strategy for CRF-Based Named Entity Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 23 publications
(21 citation statements)
references
References 29 publications
0
16
0
Order By: Relevance
“…We follow the AL settings in previous work to achieve consistent evaluation (Kim, 2020;Shelmanov et al, 2021;Liu et al, 2022). Specifically, the unlabeled pool is created by discarding labels from the original training data of each dataset; 2% of which (∼ 242 sentences) is selected for labeling at each iteration for a total of 25 iterations (examples of the first iteration are randomly sampled to serve as the seed D 0 ).…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…We follow the AL settings in previous work to achieve consistent evaluation (Kim, 2020;Shelmanov et al, 2021;Liu et al, 2022). Specifically, the unlabeled pool is created by discarding labels from the original training data of each dataset; 2% of which (∼ 242 sentences) is selected for labeling at each iteration for a total of 25 iterations (examples of the first iteration are randomly sampled to serve as the seed D 0 ).…”
Section: Discussionmentioning
confidence: 99%
“…Despite the potential of AL in reducing annotation cost for a target task, most previous AL work focuses on developing data selection strategies to maximize the model performance (Wang and Shang, 2014;Sener and Savarese, 2017;Ash et al, 2019;Kim, 2020;Liu et al, 2022;Margatina et al, 2021). As such, previous AL methods and frameworks tend to ignore the necessary time to train models and perform data selection at each AL iteration that can be significantly long and hinder annotators' productivity and model performance.…”
Section: Related Workmentioning
confidence: 99%
“…According to results from previous research, sequence level measures are superior to aggregating token-level information for sequence-labeling with CRF models (Settles and Craven, 2008;Chen et al, 2015b;Shen et al, 2017;Liu et al, 2020). We incorporate the following most representative query methods that are explored in prior work for NER tasks (Settles and Craven, 2008;Chen et al, 2015b;Shen et al, 2017;Chen et al, 2017;Siddhant and Lipton, 2018;Shelmanov et al, 2019;Chaudhary et al, 2019;Grießhaber et al, 2020;Shui et al, 2020;Ren et al, 2021;Liu et al, 2020Liu et al, , 2022Agrawal et al, 2021), in our experiments:…”
Section: Active Learning Withmentioning
confidence: 99%
“…In transfer learning, models transfer knowledge learned from data-rich languages or tasks to languages or tasks with less or no annotated data (Wang et al, 2019;Lauscher et al, 2020;Xie et al, 2018a;Yuan et al, 2019;Pires et al, 2019;Xie et al, 2018b;Plank, 2019). Active learning is an approach to maximize the utility of annotations while minimizing the annotation effort on the unlabeled target data (Chen et al, 2015a;Miller et al, 2019;Liu et al, 2020Liu et al, , 2022Chaudhary et al, 2019;Shelmanov et al, 2019;Lauscher et al, 2020). We train a German biomedical NER model building on these two approaches, addressing the following research questions: a) How to transfer knowledge from annotated English clinical narratives corpora to the German NER model?…”
Section: Introductionmentioning
confidence: 99%
“…The results showed, with the aid of AL and merely one-fourth of the training dataset, the model achieved 99% accuracy of the best deep learning models trained on the whole dataset. In (Liu, Tu, Wang, & Xu, 2020), using the BERT-CRF model, an uncertainty-based AL strategy was applied to NER and achieved satisfactory results.…”
Section: Active Learningmentioning
confidence: 99%