Proceedings of the 2019 Conference of the North 2019
DOI: 10.18653/v1/n19-1110
|View full text |Cite
|
Sign up to set email alerts
|

Cross-Topic Distributional Semantic Representations Via Unsupervised Mappings

Abstract: In traditional Distributional Semantic Models (DSMs) the multiple senses of a polysemous word are conflated into a single vector space representation. In this work, we propose a DSM that learns multiple distributional representations of a word based on different topics. First, a separate DSM is trained for each topic and then each of the topic-based DSMs is aligned to a common vector space. Our unsupervised mapping approach is motivated by the hypothesis that words preserving their relative distances in differ… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 26 publications
0
1
0
Order By: Relevance
“…Instead of using word embeddings for topic modeling, Liu et al (2015) proposed the Topical Word Embedding model which incorporates the topical information derived from standard topic models into word embedding learning by treating each topic as a pseudo-word. Briakou et al (2019) followed this route and proposed a four-stage model in which topics were first extracted from a corpus by LDA and then the topic-based word embeddings are mapped to a shared space using anchor words which were retrieved from the WordNet.…”
Section: Related Workmentioning
confidence: 99%
“…Instead of using word embeddings for topic modeling, Liu et al (2015) proposed the Topical Word Embedding model which incorporates the topical information derived from standard topic models into word embedding learning by treating each topic as a pseudo-word. Briakou et al (2019) followed this route and proposed a four-stage model in which topics were first extracted from a corpus by LDA and then the topic-based word embeddings are mapped to a shared space using anchor words which were retrieved from the WordNet.…”
Section: Related Workmentioning
confidence: 99%