Proceedings of the 15th Conference of the European Chapter of The Association for Computational Linguistics: Volume 1 2017
DOI: 10.18653/v1/e17-1119
|View full text |Cite
|
Sign up to set email alerts
|

Neural Architectures for Fine-grained Entity Type Classification

Abstract: In this work, we investigate several neural network architectures for fine-grained entity type classification and make three key contributions. Despite being a natural comparison and addition, previous work on attentive neural architectures have not considered hand-crafted features and we combine these with learnt features and establish that they complement each other. Additionally, through quantitative analysis we establish that the attention mechanism learns to attend over syntactic heads and the phrase cont… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
182
0
1

Year Published

2017
2017
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 93 publications
(183 citation statements)
references
References 22 publications
(39 reference statements)
0
182
0
1
Order By: Relevance
“…Note that we only learn the embeddings of each new entity, keeping all other Table 4: Typing Prediction: Performance on the FIGER (GOLD) dataset. Our performance is competitive with FIGER (Ling and Weld, 2012) and neural-LSTM model of Shimaoka et al (2017). Their SSIR-Full model that uses a biLSTM layer, an attention layer, combined with hand-crafted features is state-of-art for this task.…”
Section: Cold-start Entitiesmentioning
confidence: 84%
See 1 more Smart Citation
“…Note that we only learn the embeddings of each new entity, keeping all other Table 4: Typing Prediction: Performance on the FIGER (GOLD) dataset. Our performance is competitive with FIGER (Ling and Weld, 2012) and neural-LSTM model of Shimaoka et al (2017). Their SSIR-Full model that uses a biLSTM layer, an attention layer, combined with hand-crafted features is state-of-art for this task.…”
Section: Cold-start Entitiesmentioning
confidence: 84%
“…Compared to existing systems trained specifically for this task, embeddings from our approach (Model CDTE) performs competitively (see Table 4). In particular, our model performs better than the neural-LSTM model of Shimaoka et al (2017), suggesting that our multi-task linking, and typing loss facilitates effective encoding of mention contexts. Table 5: Example predictions by our models: Model CT (Ex.1) and CD (Ex.2) predict correctly when correct type prediction or background knowledge is sufficient, respectively.…”
Section: Fine-grained Typingmentioning
confidence: 90%
“…Table 1 shows performance of PthDCode on test, based on the interval [40000, 50000]; average and standard deviation are computed for 2000(20 + i), 0 ≤ i ≤ 5, as described above. PthDCode achieves clearly better results than other baseline methods -FIGER (Ling and Weld, 2012), (Yogatama et al, 2015) and (Shimaoka et al, 2017) -when trained on raw (i.e., not denoised) datasets of a similar size. Attentive encoder (Shimaoka et al, 2017) is a neural baseline for PthDCode, to which comparison in Table 1 suggests decoding of path hierarchy rather than flat classification significantly improves the performance.…”
Section: Experiments and Resultsmentioning
confidence: 90%
“…OOV vectors are randomly initialized. Similar to (Shimaoka et al, 2017), all hidden states h of the encoder-decoder were set to 100 dimension and mention lengths m to 5. Window size is w = 15.…”
Section: Experiments and Resultsmentioning
confidence: 99%
“…[12] [14] first presented a method without using feature engineering and external resources. Shimaoka et al (2016) [15] then incorporated an attention mechanism to allow the model to focus on relevant expressions.…”
Section: Related Researchesmentioning
confidence: 99%