Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1493
|View full text |Cite
|
Sign up to set email alerts
|

Language Modeling with Sparse Product of Sememe Experts

Abstract: Most language modeling methods rely on large-scale data to statistically learn the sequential patterns of words.In this paper, we argue that words are atomic language units but not necessarily atomic semantic units. Inspired by HowNet, we use sememes, the minimum semantic units in human languages, to represent the implicit semantics behind words for language modeling, named Sememe-Driven Language Model (SDLM). More specifically, to predict the next word, SDLM first estimates the sememe distribution given textu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
30
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
3
1

Relationship

3
6

Authors

Journals

citations
Cited by 30 publications
(30 citation statements)
references
References 34 publications
0
30
0
Order By: Relevance
“…HowNet, as the most well-known sememe KB, has attracted wide research attention. Previous work applies the sememe knowledge of HowNet to various NLP applications, such as word similarity computation (Liu and Li, 2002), word sense disambiguation (Gan and Wong, 2000;Zhang et al, 2005;Duan et al, 2007), sentiment analysis (Zhu et al, 2006;Dang and Zhang, 2010;Fu et al, 2013), word representation learning (Niu et al, 2017), language modeling (Gu et al, 2018), lexicon expansion (Zeng et al, 2018) and semantic rationality evaluation .…”
Section: Sememes and Hownetmentioning
confidence: 99%
See 1 more Smart Citation
“…HowNet, as the most well-known sememe KB, has attracted wide research attention. Previous work applies the sememe knowledge of HowNet to various NLP applications, such as word similarity computation (Liu and Li, 2002), word sense disambiguation (Gan and Wong, 2000;Zhang et al, 2005;Duan et al, 2007), sentiment analysis (Zhu et al, 2006;Dang and Zhang, 2010;Fu et al, 2013), word representation learning (Niu et al, 2017), language modeling (Gu et al, 2018), lexicon expansion (Zeng et al, 2018) and semantic rationality evaluation .…”
Section: Sememes and Hownetmentioning
confidence: 99%
“…HowNet (Dong and Dong, 2003) is a widely acknowledged sememe knowledge base (KB), which defines about 2,000 sememes and uses them to annotate over 100,000 Chinese words together with their English translations. Sememes and HowNet have been successfully utilized in a variety of NLP tasks including sentiment analysis (Dang and Zhang, 2010), word representation learning (Niu et al, 2017), language modeling (Gu et al, 2018), etc.…”
Section: Introductionmentioning
confidence: 99%
“…Xin et al (2018) use a similarly purposed entity typing module and a LM-enhancement module. Instead of entity type generation, Gu et al (2018) propose to explicitly decompose word generation into sememe (a semantic language unit of meaning) generation and sense generation, but requires sememe labels. Yang et al (2016) propose a pointer-network LM that can point to a 1-D or 2-D database record during inference.…”
Section: Related Workmentioning
confidence: 99%
“…Since HowNet was published (Dong and Dong, 2003), it has attracted wide attention of re-searchers. Most of related works focus on applying HowNet to specific NLP tasks (Liu and Li, 2002;Zhang et al, 2005;Sun et al, 2007;Dang and Zhang, 2010;Fu et al, 2013;Niu et al, 2017;Zeng et al, 2018;Gu et al, 2018). To the best of our knowledge, only and Jin et al (2018) conduct studies of augmenting HowNet by recommending sememes for new words.…”
Section: Related Workmentioning
confidence: 99%