Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval 2019
DOI: 10.1145/3331184.3331344
|View full text |Cite
|
Sign up to set email alerts
|

On the Effect of Low-Frequency Terms on Neural-IR Models

Abstract: Low-frequency terms are a recurring challenge for information retrieval models, especially neural IR frameworks struggle with adequately capturing infrequently observed words. While these terms are often removed from neural models -mainly as a concession to efficiency demands -they traditionally play an important role in the performance of IR models. In this paper, we analyze the effects of low-frequency terms on the performance and robustness of neural IR models. We conduct controlled experiments on three rec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
22
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
3
2

Relationship

3
5

Authors

Journals

citations
Cited by 31 publications
(27 citation statements)
references
References 14 publications
(26 reference statements)
3
22
0
Order By: Relevance
“…This suggests that relevance-matching difficulty provides positive feedback signals to the neural model to face diverse learning instances, and therefore to better generalize over different application domains. This is however true to the constraint that the vocabulary of the dataset (V ocab) is not too large so as to boost neural ranking performance as outlined in [16,36]. Looking at the residual variables (Dataset j and M j ), we can corroborate the observations made at a first glance in RQ1 regarding the model families clearly opposing (DRMM-PACRR-KNRM-VBERT) and CEDR since the former statistically exhibit higher REM metrics values than CEDR.…”
Section: Empirical Analysis Of Catastrophic Forgetting In Neuralsupporting
confidence: 73%
“…This suggests that relevance-matching difficulty provides positive feedback signals to the neural model to face diverse learning instances, and therefore to better generalize over different application domains. This is however true to the constraint that the vocabulary of the dataset (V ocab) is not too large so as to boost neural ranking performance as outlined in [16,36]. Looking at the residual variables (Dataset j and M j ), we can corroborate the observations made at a first glance in RQ1 regarding the model families clearly opposing (DRMM-PACRR-KNRM-VBERT) and CEDR since the former statistically exhibit higher REM metrics values than CEDR.…”
Section: Empirical Analysis Of Catastrophic Forgetting In Neuralsupporting
confidence: 73%
“…Nevertheless, we found this training set to produce less than optimal results and the trained BERT models show no robustness against increased re-ranking depth. This phenomena of having to tune the best re-ranking depth for effectiveness, rather than efficiency, has been studied as part of early non-BERT re-rankers [11]. With the advent of Transformer-based re-rankers, this technique became obsolete [12].…”
Section: Training Data Generationmentioning
confidence: 99%
“…In addition, we investigate classical IR models, namely BM25 [45] and RM3 PRF [1]. The BM25 and RM3 PRF models are computed using the Anserini [57] toolkit with the same setting as proposed by Hofstätter et al [24]. The results of the BM25 model is used in RM3 PRF as the first-stage retrieval.…”
Section: Experiments Setupmentioning
confidence: 99%