Proceedings of the 34th ACM/SIGAPP Symposium on Applied Computing 2019
DOI: 10.1145/3297280.3297378
|View full text |Cite
|
Sign up to set email alerts
|

Few-shot classification in named entity recognition task

Abstract: For many natural language processing (NLP) tasks the amount of annotated data is limited. This urges a need to apply semi-supervised learning techniques, such as transfer learning or meta-learning. In this work we tackle Named Entity Recognition (NER) task using Prototypical Network -a metric learning technique. It learns intermediate representations of words which cluster well into named entity classes. This property of the model allows classifying words with extremely limited number of training examples, and… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
143
0
1

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 152 publications
(144 citation statements)
references
References 15 publications
(18 reference statements)
0
143
0
1
Order By: Relevance
“…One way to train the NER model with low-resource is dictionary-based distantly supervision (Fries et al, 2017;Shang et al, 2018;Yang et al, 2018; which builds a dictionary of entities for creating training data without too much effort. Few-shot learning is another promising way for training the NER model under limited supervision by transferring prior knowledge of the source domain to a new domain (Fritzler et al, 2019;Hou et al, 2019). There are also some works that focus on redefining NER as a different problem for reducing the need of hand-labeled training data.…”
Section: Related Workmentioning
confidence: 99%
“…One way to train the NER model with low-resource is dictionary-based distantly supervision (Fries et al, 2017;Shang et al, 2018;Yang et al, 2018; which builds a dictionary of entities for creating training data without too much effort. Few-shot learning is another promising way for training the NER model under limited supervision by transferring prior knowledge of the source domain to a new domain (Fritzler et al, 2019;Hou et al, 2019). There are also some works that focus on redefining NER as a different problem for reducing the need of hand-labeled training data.…”
Section: Related Workmentioning
confidence: 99%
“…Prior work (Fritzler et al, 2019;Hou et al, 2020) on few-shot NER followed few-shot classification literature and adopted the episode evaluation methodology. Specifically, a NER system is evaluated with respect to multiple evaluation episodes.…”
Section: A Standard Evaluation Setupmentioning
confidence: 99%
“…In the context of NER, these fewshot classification methods can enable rapid building of NER systems for a new domain by labeling only a few examples per entity class. Several previous studies (Fritzler et al, 2019;Hou et al, 2020) propose using prototypical networks (Snell et al, 2017), a popular few-shot classification algorithm, to address the few-shot NER problem. However, these approaches only achieve 10 ∼ 30% F1 scores on average, when transferring knowledge between different NER datasets with one or five shot examples, warranting more effective methods for the problem.…”
Section: Introductionmentioning
confidence: 99%
“…In one of the first works on few-shot sequence labeling, Fritzler et al (2019) apply prototypical networks to few-shot named entity recognition by training a separate prototypical network for each named entity type. This design choice makes their extension of prototypical networks more restrictive than ours, which trains a single model to classify all sequence tags.…”
Section: Few-shot Learning For Sequence Labelingmentioning
confidence: 99%