Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1017
|View full text |Cite
|
Sign up to set email alerts
|

Adversarial Transfer Learning for Chinese Named Entity Recognition with Self-Attention Mechanism

Abstract: Named entity recognition (NER) is an important task in natural language processing area, which needs to determine entities boundaries and classify them into pre-defined categories. For Chinese NER task, there is only a very small amount of annotated data available. Chinese NER task and Chinese word segmentation (CWS) task have many similar word boundaries. There are also specificities in each task. However, existing methods for Chinese NER either do not exploit word boundary information from CWS or cannot filt… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
128
1
1

Year Published

2019
2019
2020
2020

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 170 publications
(138 citation statements)
references
References 24 publications
1
128
1
1
Order By: Relevance
“…Zhou et al [27] formulate NER as a joint identification to recognize entity-level features, which effectively improves performance. And Cao et al [2] also use the information of CWS for NER. Zhang et al [24] and Ding et al [5] add additional features, and the latter achieve 94.4% F1-score.…”
Section: Comparison With Previous Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Zhou et al [27] formulate NER as a joint identification to recognize entity-level features, which effectively improves performance. And Cao et al [2] also use the information of CWS for NER. Zhang et al [24] and Ding et al [5] add additional features, and the latter achieve 94.4% F1-score.…”
Section: Comparison With Previous Workmentioning
confidence: 99%
“…In order to take advantage of both character-level semantic information and word structure content, some models mix word embedding and its corresponding character vectors, and then feed mixed representation into neural network for NER [2,22,26]. The generic model mentioned above is shown in Fig.…”
Section: Introductionmentioning
confidence: 99%
“…Existing state-of-the-art systems include Peng and Dredze (2016), He and Sun (2017b), Cao et al (2018) and Zhang and Yang (2018), which leverage rich external data like cross-domain data, semi-supervised data, and lexicons, or joint-train NER and Chinese Word Segmentation (CWS). 4 In the first block of Table 2, we report the performance of the latest models.…”
Section: Weibo Datasetmentioning
confidence: 99%
“…This method allows the model to dynamically decide which source of information to use for each word, and therefore outperforming the concatenation method used in previous work. More recently, Tan et al (2018b) and Cao et al (2018) employ self-attention to directly capture the global dependencies of the inputs for NER tasks and demonstrate the effectiveness of self-attention in Chinese NER.…”
Section: Attention Mechanismmentioning
confidence: 99%
“…The use of the attention mechanism achieved great performance in machine translation by [20], setting off a new wave in the NLP field. The utilization of attention mechanism by [21] in F1-score reached 90.64% in the SIGHAN data set.…”
Section: Introductionmentioning
confidence: 97%