Proceedings of the Fifth Joint Conference on Lexical and Computational Semantics 2016
DOI: 10.18653/v1/s16-2024
|View full text |Cite
|
Sign up to set email alerts
|

Random Positive-Only Projections: PPMI-Enabled Incremental Semantic Space Construction

Abstract: We introduce positive-only projection (PoP), a new algorithm for constructing semantic spaces and word embeddings. The PoP method employs random projections. Hence, it is highly scalable and computationally efficient. In contrast to previous methods that use random projection matrices R with the expected value of 0 (i.e., E(R) = 0), the proposed method uses R with E(R) > 0. We use Kendall's τ b correlation to compute vector similarities in the resulting non-Gaussian spaces. Most importantly, since E(R) > 0, we… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2017
2017
2020
2020

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(6 citation statements)
references
References 28 publications
0
6
0
Order By: Relevance
“…For runs involving T RI, we experimented with a varying vector size from 200 to 1, 000. Moreover, we investigated (1) the initialization of the count matrix at time j with the matrix at time j − 1, (2) the contribution of positive-only projections, and (3) the application of PPMI weights, as explained in QasemiZadeh and Kallmeyer (2016). For DW 2V , we use the parameter setting proposed in Yao et al (2018).…”
Section: Methodsmentioning
confidence: 99%
“…For runs involving T RI, we experimented with a varying vector size from 200 to 1, 000. Moreover, we investigated (1) the initialization of the count matrix at time j with the matrix at time j − 1, (2) the contribution of positive-only projections, and (3) the application of PPMI weights, as explained in QasemiZadeh and Kallmeyer (2016). For DW 2V , we use the parameter setting proposed in Yao et al (2018).…”
Section: Methodsmentioning
confidence: 99%
“…I implemented the model using DyNet (Neubig et al, 2017) and Pydmrs . 6 I initialised the generative model following Emerson and Copestake (2017b) using sparse PPMI vectors (QasemiZadeh and Kallmeyer, 2016). I first trained the encoder on the initial generative model, then trained both together.…”
Section: Training Detailsmentioning
confidence: 99%
“…Random Indexing (RI) is a simple and efficient method for dimensionality reduction (Sahlgren 2005), originally used to solve clustering problems (Kaski 1998). It is also a less-travelled technique in distributional semantics (Kanerva, Kristoferson, and Holst 2000;Qasemizadeh, Kallmeyer, and Herbelot 2017;QasemiZadeh and Kallmeyer 2016). Its advocates argue that it fulfils a number of requirements of an ideal vector space construction method, in particular incrementality.…”
Section: Lexical Acquisition and The Fruit Fly Algorithmmentioning
confidence: 99%