Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing 2016
DOI: 10.18653/v1/d16-1087
|View full text |Cite
|
Sign up to set email alerts
|

Named Entity Recognition for Novel Types by Transfer Learning

Abstract: In named entity recognition, we often don't have a large in-domain training corpus or a knowledge base with adequate coverage to train a model directly. In this paper, we propose a method where, given training data in a related domain with similar (but not identical) named entity (NE) types and a small amount of in-domain training data, we use transfer learning to learn a domain-specific NE model. That is, the novelty in the task setup is that we assume not just domain mismatch, but also label mismatch.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
23
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 26 publications
(23 citation statements)
references
References 10 publications
0
23
0
Order By: Relevance
“…When considering how to deal with the situation of lack of labelled data, Pen and Yang (2010) have categorized and reviewed the reserch progress on transfer learning for classification, regression, and clustering problems. Similar to the work of Qu et al (2016) in domain adaption, in this paper we study a instance transfer strategy, which is hot today but rarely used in NER (Arnold et al, 2008;Chen et al, 2014), to make use of out-of-domain data (source domain).…”
Section: Related Workmentioning
confidence: 99%
“…When considering how to deal with the situation of lack of labelled data, Pen and Yang (2010) have categorized and reviewed the reserch progress on transfer learning for classification, regression, and clustering problems. Similar to the work of Qu et al (2016) in domain adaption, in this paper we study a instance transfer strategy, which is hot today but rarely used in NER (Arnold et al, 2008;Chen et al, 2014), to make use of out-of-domain data (source domain).…”
Section: Related Workmentioning
confidence: 99%
“…Researches were carried out on NER related TL too. Qu et al, (2016) explored TL for NER with different NE categories (different output spaces). They pre-train a linear-chain CRF on large amount annotated data in the source domain.…”
Section: Related Workmentioning
confidence: 99%
“…Given the explosion in tag-set size, they introduce automatic pruning of cross-product tags. Kim et al (2015) and Qu et al (2016) automatically learn correlations between tag-sets, given training data for both tag-sets. They rely on similar contexts for related source and target tags, such as 'professor' and 'student'.…”
Section: Related Workmentioning
confidence: 99%