Proceedings of the 2018 Conference of the North American Chapter Of the Association for Computational Linguistics: Hu 2018
DOI: 10.18653/v1/n18-1079
|View full text |Cite
|
Sign up to set email alerts
|

Nested Named Entity Recognition Revisited

Abstract: We propose a novel recurrent neural network-based approach to simultaneously handle nested named entity recognition and nested entity mention detection. The model learns a hypergraph representation for nested entities using features extracted from a recurrent neural network. In evaluations on three standard data sets, we show that our approach significantly outperforms existing state-of-the-art methods, which are feature-based. The approach is also efficient: it operates linearly in the number of tokens and th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
185
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 191 publications
(206 citation statements)
references
References 16 publications
(38 reference statements)
0
185
0
Order By: Relevance
“…In nested named entity recognition, entities can be overlapping and labeled with more than one label such as in the example "The Florida Supreme Court" containing two overlapping named entities "The Florida Supreme Court" and "Florida". 1 Recent publications on nested named entity recognition involve stacked LSTM-CRF NE recognizer (Ju et al, 2018), or a construction of a special structure that explicitly captures the nested entities, such as a constituency graph (Finkel and Manning, 2009) or various modifications of a directed hypergraph (Lu and Roth, 2015;Katiyar and Cardie, 2018;Wang and Lu, 2018).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In nested named entity recognition, entities can be overlapping and labeled with more than one label such as in the example "The Florida Supreme Court" containing two overlapping named entities "The Florida Supreme Court" and "Florida". 1 Recent publications on nested named entity recognition involve stacked LSTM-CRF NE recognizer (Ju et al, 2018), or a construction of a special structure that explicitly captures the nested entities, such as a constituency graph (Finkel and Manning, 2009) or various modifications of a directed hypergraph (Lu and Roth, 2015;Katiyar and Cardie, 2018;Wang and Lu, 2018).…”
Section: Introductionmentioning
confidence: 99%
“…Wang and Lu (2018) build a hypergraph to capture all possible entity mentions in a sentence. Katiyar and Cardie (2018) model nested entities as a directed hypergraph similar to Lu and Roth (2015), using RNNs to model the edge probabilities.…”
Section: Introductionmentioning
confidence: 99%
“…The proposed MGNER is very suitable for detecting nested named entities since every possible entity will be examined and classified. In order to validate this advantage, we compare MGNER with numerous baseline models: 1) Lu and Roth (2015) which propose the mention hypergraphs for recognizing overlapping entities; 2) Lample et al (2016) which adopt the LSTM-CRF stucture for sequence labelling; 3) Muis and Lu (2017) which introduce mention separators to tag gaps between words for recognizing overlapping mentions; 4) Xu et al (2017) that propose a local detection method; 5) Katiyar and Cardie (2018) which propose a hypergraph-based model using LSTM for learning feature representations; 6) Ju et al (2018) that use a layered model which extracts outer entities based on inner ones; 7) which propose a neural transition-based model that constructs nested mentions through a sequence of actions; 8) which adopt a neural segmental hypergraph model. Experiment results of the Nested NER task on the ACE-2004 andACE-2005 datasets are reported in Table 2.…”
Section: Resultsmentioning
confidence: 99%
“…Unlike methods such as Katiyar and Cardie (2018), it does not predict entity segmentation at each layer as discrete 0-1 labels, thus allowing the model to flexibly aggregate information across layers. Furthermore inference is greedy, without attempting to score all possible entity spans as in , which results in faster decoding (decoding requires simply a single forward pass of the network).…”
Section: Introductionmentioning
confidence: 99%