2013
DOI: 10.1007/978-3-642-36973-5_36
|View full text |Cite
|
Sign up to set email alerts
|

Two-Stage Learning to Rank for Information Retrieval

Abstract: Abstract. Current learning to rank approaches commonly focus on learning the best possible ranking function given a small fixed set of documents. This document set is often retrieved from the collection using a simple unsupervised bag-of-words method, e.g. BM25. This can potentially lead to learning a sub-optimal ranking, since many relevant documents may be excluded from the initially retrieved set. In this paper we propose a novel two-stage learning framework to address this problem. We first learn a ranking… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
36
0
1

Year Published

2015
2015
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 54 publications
(37 citation statements)
references
References 22 publications
0
36
0
1
Order By: Relevance
“…Independently of the algorithm or the loss function adopted, we observe that the cost for computing score S(q, d) is linear in the size n of the forest. As it is desirable to keep this cost as low as possible, either to comply with time budget constraints or to improve the overall effectiveness of query processing by ranking larger amount of candidate documents returned for a given query [4], we aim at reducing the complexity of a tree-based model by pruning trees. Specifically, given an input forest F providing the desired quality, CLEaVER produces a smaller forest Fp with at least the same effectiveness as F but with higher efficiency.…”
Section: Optimization Of Tree Ensemblesmentioning
confidence: 99%
“…Independently of the algorithm or the loss function adopted, we observe that the cost for computing score S(q, d) is linear in the size n of the forest. As it is desirable to keep this cost as low as possible, either to comply with time budget constraints or to improve the overall effectiveness of query processing by ranking larger amount of candidate documents returned for a given query [4], we aim at reducing the complexity of a tree-based model by pruning trees. Specifically, given an input forest F providing the desired quality, CLEaVER produces a smaller forest Fp with at least the same effectiveness as F but with higher efficiency.…”
Section: Optimization Of Tree Ensemblesmentioning
confidence: 99%
“…In (Lai et al 2013), the authors presented a sparse learning-to-rank model for information retrieval. Dang et al (2013) proposed a two-stage learning-to-rank framework to address the problem of sub-optimal ranking when many relevant documents are excluded from the ranking list using bag-of-words retrieval models. In (Tan et al 2013), the authors proposed a model which directly optimizes the ranking measure without resorting to any upper bounds or approximations.…”
Section: Related Workmentioning
confidence: 99%
“…For our purposes, type(s 6 ) = visible, as we can substitute f 6 (s 2 , s 3 , p 7 ) in place of s 6 into the input of f 8 . Note that this can work from the other end as well: given that s 8 = f 8 (s 5 , s 6 , p 9 , p 10 ), given the substitution of s 6 and that type(s 8 ) = visible, we can then construct an inverse f −1 8 to recover what the value of s 5 is:…”
Section: Inferring Hidden Subscoresmentioning
confidence: 99%
“…In this work, we present a parameter tuning approach which can be used for either stage. This approach is especially impactful for the retrieval stage -while a large amount of work focuses on optimizing the parameters of the ranking stage, relatively little work [3,8] covers parameter tuning at the retrieval stage.…”
Section: Introductionmentioning
confidence: 99%