2016 International Conference on Big Data and Smart Computing (BigComp) 2016
DOI: 10.1109/bigcomp.2016.7425938
|View full text |Cite
|
Sign up to set email alerts
|

Bagging-based active learning model for named entity recognition with distant supervision

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
4

Relationship

1
7

Authors

Journals

citations
Cited by 10 publications
(3 citation statements)
references
References 7 publications
0
3
0
Order By: Relevance
“…To construct NER's training data automatically, Jingbo et al applied named-entity dictionary refinement and developed a deep-learning-based NER by mapping the result [8]. In [9], the initial NER training data were automatically constructed through dictionary mapping of named entities and raw data from Wikipedia. The study reduced the number of errors in the initial NER training data by applying active learning with a bagging method and a conditional random field (CRF) method; thus, the NER system's performance was improved.…”
Section: A Automatic Construction Of Training Data Using Distant Supmentioning
confidence: 99%
“…To construct NER's training data automatically, Jingbo et al applied named-entity dictionary refinement and developed a deep-learning-based NER by mapping the result [8]. In [9], the initial NER training data were automatically constructed through dictionary mapping of named entities and raw data from Wikipedia. The study reduced the number of errors in the initial NER training data by applying active learning with a bagging method and a conditional random field (CRF) method; thus, the NER system's performance was improved.…”
Section: A Automatic Construction Of Training Data Using Distant Supmentioning
confidence: 99%
“…Furthermore, we use a bagging-based active learning process to refine noises in the corpus (i.e., words tannotated with incorrect NE classes), which refines the NER accuracy of our system. A preliminary discussion of our model was presented in [19] as a short paper. The preliminary model did not consider that most of entry words in a NE dictionary can have multiple NE classes during the distant supervision phase.…”
Section: Previous Workmentioning
confidence: 99%
“…Previous studies of Korean NER systems focused on the morphological-level text that is made using morphological analysis and handcrafted features (Na et al 2019;Yu and Ko 2017;Lee et al 2016) while there are various studies about character-level tagger (Kuru, Can, and Yuret 2016) or using sub-character information in Chinese (Dong et al 2016). Although morphological-level NER has advantages in utilizing richer linguistic information such as part-of-speech (POS) tags and word boundaries, it suffers from the potential issue of error propagation that error of morphological analysis results commonly lead to NER errors.…”
Section: Introductionmentioning
confidence: 99%