Lecture Notes in Computer Science
DOI: 10.1007/978-3-540-78646-7_8
|View full text |Cite
|
Sign up to set email alerts
|

Effective Pre-retrieval Query Performance Prediction Using Similarity and Variability Evidence

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
180
0
1

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 111 publications
(183 citation statements)
references
References 13 publications
2
180
0
1
Order By: Relevance
“…Later, Scholer et al [5] reported that using the maximum IDF value of any term in a query gives the best correlation on the TREC web data. These results were confirmed and extended to other TREC collections [7].…”
Section: Related Worksupporting
confidence: 77%
See 2 more Smart Citations
“…Later, Scholer et al [5] reported that using the maximum IDF value of any term in a query gives the best correlation on the TREC web data. These results were confirmed and extended to other TREC collections [7].…”
Section: Related Worksupporting
confidence: 77%
“…More recently, Zhao et al [7] presented two families of pre-retrieval predictors. The first is based on the similarity between a query and the overall document collection, the second focuses on the variability in how query terms are distributed across documents.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The features used to train this classifier included query performance estimators (Average Inverse Collection Term Frequency (AvICTF) [12], Simplified Clarity Score (SCS)) [11], the derivates of the similarity score between collection and query (SumSCQ, AvSCQ, MaxSCQ) [27], result set size and the un-normalised BM25 document scores for the top five documents. Their classifier achieved 78 % accuracy on the FAQ SMS training data using a leave-one-out validation.…”
Section: Related Workmentioning
confidence: 99%
“…Seven of these predictors were pre-retrieval predictors and these were : Average Pointwise Mutual Information (AvPMI) [10], Simplified Clarity Score (SCS) [11], Average Inverse Collection Term Frequency (AvICTF) [12], Average Inverse Document Frequency (AvIDF) [10] and the derivatives of the similarity score between collection and query (SumSCQ, AvSCQ, MaxSCQ) [27]. One post-retrieval predictor was used, the Clarity Score (CS) [5].…”
Section: Creating Training and Testing Instances For Missing Content mentioning
confidence: 99%