Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1482
|View full text |Cite
|
Sign up to set email alerts
|

Word Mover’s Embedding: From Word2Vec to Document Embedding

Abstract: While the celebrated Word2Vec technique yields semantically rich representations for individual words, there has been relatively less success in extending to generate unsupervised sentences or documents embeddings. Recent work has demonstrated that a distance measure between documents called Word Mover's Distance (WMD) that aligns semantically similar words, yields unprecedented KNN classification accuracy. However, WMD is expensive to compute, and it is hard to extend its use beyond a KNN classifier. In this … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
88
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
4
1

Relationship

1
9

Authors

Journals

citations
Cited by 85 publications
(88 citation statements)
references
References 46 publications
0
88
0
Order By: Relevance
“…It also reveals that probably there is still room to improve ranking performance, especially for cases in which a larger number of seed websites is available. Given recent advances in text representation using dense vector representations and document distance computation in the embedding space [17,33], this is a promising area for future work. While this is a limitation of our ranking approach, it is still suitable for our context, where we assume to have only a small list of seeds as input.…”
Section: Website Ranking Evaluationmentioning
confidence: 99%
“…It also reveals that probably there is still room to improve ranking performance, especially for cases in which a larger number of seed websites is available. Given recent advances in text representation using dense vector representations and document distance computation in the embedding space [17,33], this is a promising area for future work. While this is a limitation of our ranking approach, it is still suitable for our context, where we assume to have only a small list of seeds as input.…”
Section: Website Ranking Evaluationmentioning
confidence: 99%
“…In this work, instead of using Random Features to approximate a pre-defined kernel function, we overcome all these aforementioned issues by generalizing Random Features to develop a new family of efficient and effective string kernels that not only are positive-definite but also reduce the computational complexity from quadratic to linear in both the number and the length of strings. Note that, our approach is different from a recent work [45] on distance kernel learning that mainly focuses on theoretical analysis of these kernels on structured data like time-series [46] and text [44]. Instead, we focus on developing empirical methods that could often outperform or are highly competitive to other state-of-the-art approaches, including kernel based and Recurrent Neural Networks based methods, as we will show in our experiments.…”
Section: Conventional Random Features For Scaling Up Kernel Machinementioning
confidence: 99%
“…Le and Mikolov (2014); Li et al (2015); Dai et al (2015) explored Paragraph Vector with various lengths (sentence, paragraph, document) trained on next word/n-gram prediction given context sampled from the paragraph. The work from Roy et al (2016); Chen (2017); Wu et al (2018) obtained document embeddings from word-level embeddings. More recent work has been focused on learning document embeddings through hierarchical training.…”
Section: Related Workmentioning
confidence: 99%