Proceedings of the 2nd Workshop on Representation Learning for NLP 2017
DOI: 10.18653/v1/w17-2630
|View full text |Cite
|
Sign up to set email alerts
|

Deep Active Learning for Named Entity Recognition

Abstract: Deep neural networks have advanced the state of the art in named entity recognition. However, under typical training procedures, advantages over classical methods emerge only with large datasets. As a result, deep learning is employed only when large public datasets or a large budget for manually labeling data is available. In this work, we show that by combining deep learning with active learning, we can outperform classical methods even with a significantly smaller amount of training data.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
275
2
2

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 308 publications
(280 citation statements)
references
References 28 publications
1
275
2
2
Order By: Relevance
“…The dataset contains several sources of noise, including OCR errors, referencing errors or inconsistencies, annotation errors. In part for this reason, we consider as the most important next step for future work to explore how active learning or semi-supervised learning techniques might be used in order to maximize the model gain while at the same time minimizing the costly process of manual annotation (Peters et al, 2017;Shen et al, 2018). At the same time, we plan to explore how to align and use existing annotated datasets with coverage in the arts and humanities (Anzaroot and McCallum, 2013).…”
Section: Resultsmentioning
confidence: 99%
“…The dataset contains several sources of noise, including OCR errors, referencing errors or inconsistencies, annotation errors. In part for this reason, we consider as the most important next step for future work to explore how active learning or semi-supervised learning techniques might be used in order to maximize the model gain while at the same time minimizing the costly process of manual annotation (Peters et al, 2017;Shen et al, 2018). At the same time, we plan to explore how to align and use existing annotated datasets with coverage in the arts and humanities (Anzaroot and McCallum, 2013).…”
Section: Resultsmentioning
confidence: 99%
“…To the best of our knowledge, active learning has not been applied to classification tasks for scientific text yet. Recent publications demonstrate the effectiveness of active learning for NLP tasks such as Named Entity Recognition (NER) [37] and sentence classification [44]. Siddhant and Lipton [38] and Shen et.…”
Section: Scientific Corporamentioning
confidence: 99%
“…al. [37] compare several sampling strategies on NLP tasks and show that Maximum Normalized Log-Probability (MNLP) based on uncertainty sampling performs well in NER.…”
Section: Scientific Corporamentioning
confidence: 99%
“…OntoNotes 5.0. On OntoNotes 5.0, we compare the proposed GRN with NER models that also reported performance on it, including (Chiu and Nichols 2016;Shen et al Model Mean(±std) F 1 Mean P/R (Chiu and Nichols 2016) 86.28 ± 0.26 86.04 / 86.53 (Shen et al 2017) 86.63 ± 0.49 (Durrett and Klein 2014) 84.04 85.22 / 82.89 (Passos, Kumar, and McCallum 2014) 82.30 (? ) 83.45 CNN-BiLSTM-Att-CRF 87.25±0.17 ID-CNN (Strubell et al 2017) 86.84 ± 0.19 GRN 87.67 ± 0.17 87.79 / 87.56 Table 3, GRN can obtain the state-of-the-art NER performance on OntoNotes 5.0, which further demonstrates its effectiveness.…”
Section: Performance Comparisonmentioning
confidence: 99%