Proceedings of the 24th ACM International on Conference on Information and Knowledge Management 2015
DOI: 10.1145/2806416.2806475
|View full text |Cite
|
Sign up to set email alerts
|

Short Text Similarity with Word Embeddings

Abstract: Determining semantic similarity between texts is important in many tasks in information retrieval such as search, query suggestion, automatic summarization and image finding. Many approaches have been suggested, based on lexical matching, handcrafted patterns, syntactic parse trees, external sources of structured semantic knowledge and distributional semantics. However, lexical features, like string matching, do not capture semantic similarity beyond a trivial level. Furthermore, handcrafted patterns and exter… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
240
0
4

Year Published

2017
2017
2021
2021

Publication Types

Select...
5
4
1

Relationship

2
8

Authors

Journals

citations
Cited by 355 publications
(252 citation statements)
references
References 23 publications
2
240
0
4
Order By: Relevance
“…In this survey, we review neural models for textual similarity only if the model is evaluated for retrieval of similar textual units. Limiting to Similar Item Retrieval, we exclude works on neural models for general purpose textual similarity such as Hill et al (2016), Kenter and de Rijke (2015).…”
mentioning
confidence: 99%
“…In this survey, we review neural models for textual similarity only if the model is evaluated for retrieval of similar textual units. Limiting to Similar Item Retrieval, we exclude works on neural models for general purpose textual similarity such as Hill et al (2016), Kenter and de Rijke (2015).…”
mentioning
confidence: 99%
“…First, using pre-trained word embeddings like combining traditional retrieval models with an embedding-based translation model [16,58], using pre-trained embeddings for query expansion to improve retrieval [57], and representing documents as Bag-of-Word-Embeddings (BoWE) [20,27]. Second, learning representations from scratch like learning representations of words and documents [28,32] and employing them in retrieval task [2,3], and learning representations in an end-to-end neural model for learning a speci c task like entity ranking for expert nding [53] or product search [52].…”
Section: Objectivesmentioning
confidence: 99%
“…Although research on natural language processing and text mining can be regarded as mature, it has largely focused on the assumption of well-written "long enough" documents [36]. Several methods for word and phrase similarity have been proposed in recent years [38,39], which aim at measuring similarity between text that may not contain any words in common. While not all these methods are directly applicable for social media, they represent promising approaches to identify related text content.…”
Section: Page 766mentioning
confidence: 99%