2021
DOI: 10.3390/sym13050786
|View full text |Cite
|
Sign up to set email alerts
|

Low-Resource Named Entity Recognition via the Pre-Training Model

Abstract: Named entity recognition (NER) is an important task in the processing of natural language, which needs to determine entity boundaries and classify them into pre-defined categories. For low-resource languages, most state-of-the-art systems require tens of thousands of annotated sentences to obtain high performance. However, there is minimal annotated data available about Uyghur and Hungarian (UH languages) NER tasks. There are also specificities in each task—differences in words and word order across languages … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(11 citation statements)
references
References 27 publications
0
4
0
Order By: Relevance
“…Currently, due to the increasing influence of pre-training models such as the Bidirectional Encoder Representations from Transformers (BERT) in natural language processing (NLP) research, another study ( Chen et al, 2021 ) introduced pre-training models in NER research to enhance NER models through the powerful semantic representation ability of pre-training models for semantic understanding of text, thus achieving better entity recognition results.…”
Section: Related Workmentioning
confidence: 99%
“…Currently, due to the increasing influence of pre-training models such as the Bidirectional Encoder Representations from Transformers (BERT) in natural language processing (NLP) research, another study ( Chen et al, 2021 ) introduced pre-training models in NER research to enhance NER models through the powerful semantic representation ability of pre-training models for semantic understanding of text, thus achieving better entity recognition results.…”
Section: Related Workmentioning
confidence: 99%
“…The advent of cross-lingual models has been at the aid of many downstream tasks plagued with low-resourcedness such as document classification 48 , POS tagging 20 , sentiment analyses 49 , and named entity recognition 50 . The primary goal of these models is to learn transferable language-generic knowledge encoded inside sound embedding spaces(of high-resourced languages) obtained from large-enough language representation and inject this mined "knowledge" into the embedding spaces of low-resourced languages to use for downstream tasks.…”
Section: Related Workmentioning
confidence: 99%
“…These IE applications include sequence tagging tasks such as Named-Entity Recognition (NER) and Part-of-Speech (POS) tagging. N NER is a task that processes natural language, classifying and grouping, for example, words into categories (also known as phrase types) [20]. With the advent of big data and large datasets, classifying natural language in these datasets has become increasingly important.…”
Section: Introductionmentioning
confidence: 99%
“…tions are able to apply NER in customer support, content classification, and search and recommendation engines [21]. Furthermore, NER findings may be transferred to other NLP tasks such as MT, automatic text summarization, and knowledge base construction [20]. Lack of data severely impedes performance on NER tasks with low-resourced languages [20].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation