Proceedings of the 34th International ACM SIGIR Conference on Research and Development in Information Retrieval 2011
DOI: 10.1145/2009916.2010039
|View full text |Cite
|
Sign up to set email alerts
|

Repeatable and reliable search system evaluation using crowdsourcing

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
67
0

Year Published

2012
2012
2013
2013

Publication Types

Select...
8

Relationship

1
7

Authors

Journals

citations
Cited by 82 publications
(76 citation statements)
references
References 15 publications
5
67
0
Order By: Relevance
“…While LTR offers high performance, it critically depends on the availability of relevance judgments for training. We observed from our experiments based on real users (via a crowd sourcing based evaluation recently proposed in [3]) that the final results strongly correlate with the number of visits (#visits) that is captured in the access logs. We provide a detailed analysis of this correlation and for the case where training data and ground truth is not easy to obtain, we propose the use of #visits as an alternative.…”
Section: Introductionmentioning
confidence: 80%
See 2 more Smart Citations
“…While LTR offers high performance, it critically depends on the availability of relevance judgments for training. We observed from our experiments based on real users (via a crowd sourcing based evaluation recently proposed in [3]) that the final results strongly correlate with the number of visits (#visits) that is captured in the access logs. We provide a detailed analysis of this correlation and for the case where training data and ground truth is not easy to obtain, we propose the use of #visits as an alternative.…”
Section: Introductionmentioning
confidence: 80%
“…As performance measures, we use the standard measures NDCG and Spearman's correlation coefficient. We build upon the data, queries and methodology proposed by the recent SemSearch Challenge evaluation initiative [3] …”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Noteworthy, one source (MS) is dominant in the publication scenario and is part of 83% of the co-references. We followed the methodology of [15] to obtain relevance judgments for the ranking evaluation. We rated the top-10 results for each query.…”
Section: Methodsmentioning
confidence: 99%
“…The fact that the pool will have versions assessed by different judges over time is not a problem. The ranking between the judged systems will be the same as if judges would have assessed all documents in the same day [29].…”
Section: Reusabilitymentioning
confidence: 99%