2018
DOI: 10.1007/978-3-319-73606-8_4
|View full text |Cite
|
Sign up to set email alerts
|

From Vector Space Models to Vector Space Models of Semantics

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(3 citation statements)
references
References 5 publications
0
3
0
Order By: Relevance
“…We just count the frequency of each word in the piece of text and created a dictionary of them which is called tokenization process in NLP which is then passed to countvectorize object in scikit learn package to create a set of maximum features. We use fit transform method to model (Ganesh et al, 2016) the bag of words feature vector which are stored in an array.…”
Section: Importing Training and Cross-validation Frommentioning
confidence: 99%
“…We just count the frequency of each word in the piece of text and created a dictionary of them which is called tokenization process in NLP which is then passed to countvectorize object in scikit learn package to create a set of maximum features. We use fit transform method to model (Ganesh et al, 2016) the bag of words feature vector which are stored in an array.…”
Section: Importing Training and Cross-validation Frommentioning
confidence: 99%
“…We used Random Kitchen Sink (RKS) (Sathyan et al, 2018) algorithms with character word-bound based Term Frequency-Inverse Document Frequency (TF-IDF) (Barathi Ganesh et al, 2016) for text representation and classification was performed using Support Vector Machines (SVM) classifier (Soman et al, 2009), (Premjith et al, 2019). The rest of the paper is organised as follows: Section 2 describes about the related works, Section 3 describes about the Datasets, Section 4 describes about the preprocessing and different methods used, Section 5 describes about the result and analysis and Section 6 concludes the paper.…”
Section: Introductionmentioning
confidence: 99%
“…One of most naive way of representing word in vector form is one hot representation but it is very ineffective way for representing words in a large corpus since the length of one hot vector grows as the vocabulary increases, so we need a better and more effective way which captures some semantic similarities (Ganesh et al, 2016) between nearby words, thus creating the representation for words bring beneficial info about the word and its actual meaning, the methods which encodes these information about the words are called word embedding models, they are categorized into count based and predictive word embedding models. Both embedding models at least some way share syntactic meaning .…”
Section: Related Workmentioning
confidence: 99%