2019
DOI: 10.1007/s10339-019-00934-x
|View full text |Cite
|
Sign up to set email alerts
|

Bridging the theoretical gap between semantic representation models without the pressure of a ranking: some lessons learnt from LSA

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
11
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2

Relationship

2
7

Authors

Journals

citations
Cited by 17 publications
(11 citation statements)
references
References 114 publications
0
11
0
Order By: Relevance
“…These models formally represent the semantic meaning of words as coordinates in a high‐dimensional vector space. In general, vector space models provide high‐quality input to emulate different cognitive mechanisms (see Günther et al., 2019; Jones, Gruenenfelder, & Recchia, 2018; or Jorge‐Botana, Olmos, & Luzón, 2020 for a recent revision on the use of these models to simulate cognitive process). Given that the semantic information of vector space models is derived from co‐occurrences of written words in texts (usually from a corpus of tens of thousands of documents), their semantic representations are considered amodal.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…These models formally represent the semantic meaning of words as coordinates in a high‐dimensional vector space. In general, vector space models provide high‐quality input to emulate different cognitive mechanisms (see Günther et al., 2019; Jones, Gruenenfelder, & Recchia, 2018; or Jorge‐Botana, Olmos, & Luzón, 2020 for a recent revision on the use of these models to simulate cognitive process). Given that the semantic information of vector space models is derived from co‐occurrences of written words in texts (usually from a corpus of tens of thousands of documents), their semantic representations are considered amodal.…”
Section: Introductionmentioning
confidence: 99%
“…The methodology that uses the representations of LSA to generate different vector spaces representing different stages of vocabulary development is known as Word Maturity (Kireyev & Landauer, 2011; Landauer et al., 2011). LSA is well suited for this methodology because all words and texts are expressed on an orthogonal basis and this makes it possible to keep the similarity distances even with the transformations performed by the procedure (see Jorge‐Botana et al., 2020 for a discussion of the advantages of orthogonality). A brief explanation on how Word Maturity works will be given now.…”
Section: Introductionmentioning
confidence: 99%
“…To build the topic model of LSA, "stop words" such as "the," "is," and "at" must be removed [18]. ese stop words have little substantive meaning for the topic model.…”
Section: Design Of Data Processing Modulementioning
confidence: 99%
“…Vector techniques are based on the automatic processing of large text corpora representing a language of a general or specific domain, although some studies recommend specific domain corpora (Kwantes et al, 2016). There are different computational models in which word occurrences are algebraically vectorized such as LSA, word2vec, or BEAGLE (for a revision on space models see Günther, Rinaldi, & Marelli, 2019;Jorge-Botana, Olmos, & Luzón, 2020;Jones, Willits, & Dennis, 2015;or McNamara, 2011). All of them coincide in that they represent the lexicon in a reduced dimensionality vector space.…”
Section: Introductionmentioning
confidence: 99%