2016
DOI: 10.48550/arxiv.1606.04640
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Siamese CBOW: Optimizing Word Embeddings for Sentence Representations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
24
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 19 publications
(24 citation statements)
references
References 0 publications
0
24
0
Order By: Relevance
“…The dense vector representation of text can be constructed by the ensemble of word embeddings (Mikolov et al, 2013) in the text. The Siamese CBOW model (Kenter et al, 2016) constructs the sentence embedding by averaging the word embeddings and uses the embedding similarities among the sentence, its adjacent sentences and randomly chosen sentences as training target to fine tune the sentence embedding. The Word Mover's Distance (WMD) (Kusner et al, 2015) measures the similarity of two texts by calculating the minimum accumulate distance from all the embedded words in one text to the embedded words in the other text.…”
Section: Related Workmentioning
confidence: 99%
“…The dense vector representation of text can be constructed by the ensemble of word embeddings (Mikolov et al, 2013) in the text. The Siamese CBOW model (Kenter et al, 2016) constructs the sentence embedding by averaging the word embeddings and uses the embedding similarities among the sentence, its adjacent sentences and randomly chosen sentences as training target to fine tune the sentence embedding. The Word Mover's Distance (WMD) (Kusner et al, 2015) measures the similarity of two texts by calculating the minimum accumulate distance from all the embedded words in one text to the embedded words in the other text.…”
Section: Related Workmentioning
confidence: 99%
“…Siamese C-BOW (Kenter et al, 2016) shares a common concept with SIF and Sent2vec: defining a sentence vector as the average of word embedding vectors.…”
Section: Related Workmentioning
confidence: 99%
“…Assembling successful distributional word representations (for example, GloVe (Pennington et al, 2014)) into sentence representations is an active research topic. Different from previous studies (for example, doc2vec (Mikolov et al, 2013), skip-thought vectors , Siamese CBOW (Kenter et al, 2016)), our main contribution is to represent sentences using non-vector space representations: a sentence can be well represented by the subspace spanned by the context word vectors -such a method naturally builds on any word representation method. Due to the widespread use of word2vec and GloVe, we use their publicly available word representations -word2vec (Mikolov et al, 2013) trained using Google News 1 and GloVe (Pennington et al, 2014) trained using Common Crawl 2 -to test our observations.…”
Section: Geometry Of Sentencesmentioning
confidence: 99%
“…A sentence contains rich syntactic information and can be modeled through sophisticated neural networks (e.g., convolutional neural networks (Kim, 2014;Kalchbrenner et al, 2014), recurrent neural networks (Sutskever et al, 2014;Le and Mikolov, 2014;Hill et al, 2016) and recursive neural networks (Socher et al, 2013)). Another simple and common approach ignores the latent structure of sentences: a prototypical approach is to represent a sentence by summing or averaging over the vectors of the words in this sentence (Wieting et al, 2015;Adi et al, 2016;Kenter et al, 2016). Recently, Wieting et al (2015); Adi et al (2016) reveal that even though the latter approach ignores all syntactic information, it is simple, straightforward, and remarkably robust at capturing the sentential semantics.…”
Section: Introductionmentioning
confidence: 99%