2020
DOI: 10.3390/app10217557
|View full text |Cite
|
Sign up to set email alerts
|

Delayed Combination of Feature Embedding in Bidirectional LSTM CRF for NER

Abstract: Named Entity Recognition (NER) plays a vital role in natural language processing (NLP). Currently, deep neural network models have achieved significant success in NER. Recent advances in NER systems have introduced various feature selections to identify appropriate representations and handle Out-Of-the-Vocabulary (OOV) words. After selecting the features, they are all concatenated at the embedding layer before being fed into a model to label the input sequences. However, when concatenating the features, inform… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 11 publications
(5 citation statements)
references
References 34 publications
0
4
0
Order By: Relevance
“…From recent literature, there has been little research on information extraction systems in Ethiopian and foreign languages [3], [7], [8], [10]. Because of its grammatical nuances, syntactic structure, lexical structure, writing system, morphological structure, and encoding style, existing approaches created for other languages are difficult to apply directly to the Ge'ez language [21].…”
Section: Related Workmentioning
confidence: 99%
“…From recent literature, there has been little research on information extraction systems in Ethiopian and foreign languages [3], [7], [8], [10]. Because of its grammatical nuances, syntactic structure, lexical structure, writing system, morphological structure, and encoding style, existing approaches created for other languages are difficult to apply directly to the Ge'ez language [21].…”
Section: Related Workmentioning
confidence: 99%
“…Unlike a traditional LSTM, the BiLSTM can run input sequence in two ways: forward layer from past to future and backward layer from future to past. By this approach, two hidden states from two LSTM models are combined with preserving information from both the past and future [38], which improve the long-term learning dependencies and the model's accuracy.…”
Section: Proposed Model Architecturementioning
confidence: 99%
“…More recently, NN-based models have attained a state-of-the-art status in NER [6,15,22,23] and other sequence tagging tasks [8,24]. These models primarily rely on the feature representation at character-level [10,11], word-level [25,26], and joint character-level and word-level [9,13] to encode a word boundary and its context.…”
Section: Agglutinative and Inflectional Morphologymentioning
confidence: 99%
“…The limitation of the pretrained word embedding model is the representation of a word in a single embedding without encoding its context [23].To tackle this drawback, two different word representation techniques have been developed: one is pattern-based wordlevel encoding, and the other is LSTM-based character-level word encoding. The former (word-level) approach makes it possible to learn pattern-based word representations by encoding them with different literals but with the same representations [8,9].…”
Section: Task-specific Contextual Representation Learningmentioning
confidence: 99%