2021
DOI: 10.1007/978-3-030-72113-8_31
|View full text |Cite
|
Sign up to set email alerts
|

CEQE: Contextualized Embeddings for Query Expansion

Abstract: In this work we leverage recent advances in context-sensitive language models to improve the task of query expansion. Contextualized word representation models, such as ELMo and BERT, are rapidly replacing static embedding models. We propose a new model, Contextualized Embeddings for Query Expansion (CEQE), that utilizes queryfocused contextualized embedding vectors. We study the behavior of contextual representations generated for query expansion in ad-hoc document retrieval. We conduct our experiments on pro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
32
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 24 publications
(32 citation statements)
references
References 28 publications
0
32
0
Order By: Relevance
“…Other approaches such as DeepCT [9] and doc2query [24,25] use neural models to augment documents before indexing using a traditional inverted index. CEQE [23] generates words to expand the initial query, which is then executed on the inverted index. However, returning the BERT embeddings back to textual word forms can result in polysemous words negatively affecting retrieval.…”
Section: Discussionmentioning
confidence: 99%
See 3 more Smart Citations
“…Other approaches such as DeepCT [9] and doc2query [24,25] use neural models to augment documents before indexing using a traditional inverted index. CEQE [23] generates words to expand the initial query, which is then executed on the inverted index. However, returning the BERT embeddings back to textual word forms can result in polysemous words negatively affecting retrieval.…”
Section: Discussionmentioning
confidence: 99%
“…However the use of BERT models directly within the pseudo-relevance feedback mechanism has seen comparatively little use in the literature. The current approaches leveraging the BERT contextualised embeddings for PRF are Neural PRF [18], BERT-QE [37] and CEQE [23].…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…It is clear that integrating PRF signals into deep language models implies a trade-off between effectiveness and efficiency. While current approaches ignored efficiency, the majority still achieved marginal improvements in effective- maintaining efficiency: (i) by concatenating the feedback passages with the original query to form the new queries that contain the relevant signals, (ii) by pre-generating passage collection embeddings and performing PRF in the vector space, because embeddings promise to capture the semantic similarity between terms [10,13,22,36,37,45,56,57],…”
Section: Related Workmentioning
confidence: 99%