Proceedings of the 31st ACM International Conference on Information &Amp; Knowledge Management 2022
DOI: 10.1145/3511808.3557367
|View full text |Cite
|
Sign up to set email alerts
|

Introducing Neural Bag of Whole-Words with ColBERTer

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 19 publications
(12 citation statements)
references
References 27 publications
0
6
0
Order By: Relevance
“…• ANCE [53] and ADORE [58]: two effective dense retrieval models based on BERT-Base [13] that use the model itself to mine hard negative documents. • RocketQA [37], Margin-MSE [17], and TAS-B [18]: effective dense retrieval models that use knowledge distillation from a BERT reranking model (a cross-encoder) in addition to various techniques for negative sampling. • Contriever-FT [20]: a single vector dense retrieval model that is pre-trained for retrieval tasks and then fine-tuned on MS MARCO.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…• ANCE [53] and ADORE [58]: two effective dense retrieval models based on BERT-Base [13] that use the model itself to mine hard negative documents. • RocketQA [37], Margin-MSE [17], and TAS-B [18]: effective dense retrieval models that use knowledge distillation from a BERT reranking model (a cross-encoder) in addition to various techniques for negative sampling. • Contriever-FT [20]: a single vector dense retrieval model that is pre-trained for retrieval tasks and then fine-tuned on MS MARCO.…”
Section: Resultsmentioning
confidence: 99%
“…Existing single vector dense retrieval models uses a 𝑘-dimensional latent vector to represent each query or each query token [17,23,53,57]. We argue that these dense retrieval models can benefit from modeling uncertainty in representation learning.…”
Section: The Mrl Frameworkmentioning
confidence: 99%
See 1 more Smart Citation
“…Our baselines for comparison are the original OKAPI BM25 algorithm [24] and the BERT-based method of [54] denoted by BERT-rank. Our vector matching method is denoted by VM, and we use all varieties of text data (described in Section 2.2) and text representation vectors (described in Section 2.3).…”
Section: Methodsmentioning
confidence: 99%
“…This paper uses the loss of SPLADE with a combination that delivers the best result in our training process. 𝐿 𝑅 is the ranking loss with margin MSE for knowledge distillation [12]. [37], DeepCT [5], DeepImpact [23], and uniCOIL [10,20].…”
Section: Introductionmentioning
confidence: 99%