Lecture Notes in Computer Science
DOI: 10.1007/978-3-540-78646-7_5
|View full text |Cite
|
Sign up to set email alerts
|

Here or There

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
46
0

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 72 publications
(53 citation statements)
references
References 11 publications
1
46
0
Order By: Relevance
“…Note that with TABS, for a set of k search results, we collect one preference judgment out of a total of k(k−1)/2 possible judgments. We do so to make the human annotation task very light; pairwise preference judgments are easier to obtain from human assessors than requiring them to consistently rank a larger set of documents [3], [5]. Below we explain how we can use this highly incomplete set of preference judgments to evaluate ranking strategies.…”
Section: Evaluation Of Ranking Measures a Preference Judgments mentioning
confidence: 99%
See 2 more Smart Citations
“…Note that with TABS, for a set of k search results, we collect one preference judgment out of a total of k(k−1)/2 possible judgments. We do so to make the human annotation task very light; pairwise preference judgments are easier to obtain from human assessors than requiring them to consistently rank a larger set of documents [3], [5]. Below we explain how we can use this highly incomplete set of preference judgments to evaluate ranking strategies.…”
Section: Evaluation Of Ranking Measures a Preference Judgments mentioning
confidence: 99%
“…Though the definition above is similar in spirit to the metrics proposed in [1], [2], [3], it is different in its computation of the accuracy of a ranking strategy. In all these previously proposed evaluation measures, more than one preference judgment per query is considered.…”
Section: Definition 8 (Ranking Accuracy)mentioning
confidence: 99%
See 1 more Smart Citation
“…Now, our key idea is to design a preference-based measure to score each ranked list by treating these inferred incomplete preference relations between documents as our golden standard. In this study we use precision of preference(ppref) [4].…”
Section: Methodsmentioning
confidence: 99%
“…It can be a challenging task for humans to produce graded coherence assessments of topics. Therefore, we apply a pairwise preference user study [22] to gather human judgments. A similar method has been previously used to compare summarisation algorithms [23].…”
Section: Comparison Of Coherence Metricsmentioning
confidence: 99%