2000
DOI: 10.1002/(sici)1097-4571(2000)51:9<834::aid-asi60>3.0.co;2-1
|View full text |Cite
|
Sign up to set email alerts
|

When information retrieval measures agree about the relative quality of document rankings

Abstract: The variety of performance measures available for information retrieval systems, search engines, and network filtering agents can be confusing to both practitioners and scholars. Most discussions about these measures address their theoretical foundations and the characteristics of a measure that make it desirable for a particular application. In this work, we consider how measures of performance at a point in a search may be formally compared. Criteria are developed that allow one to determine the percent of t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
14
0

Year Published

2000
2000
2013
2013

Publication Types

Select...
5
2
1

Relationship

4
4

Authors

Journals

citations
Cited by 27 publications
(15 citation statements)
references
References 19 publications
0
14
0
Order By: Relevance
“…Other performance measures could be used for such a study, or the relationship between performance with this measure and performance with other measures might be studied (Losee, 2000). An earlier study on distributed information retrieval provides methods using Average Search Length as a performance measure that can address issues such as the clustering problems discussed above (Losee & Church, 2004).…”
Section: Discussionmentioning
confidence: 99%
“…Other performance measures could be used for such a study, or the relationship between performance with this measure and performance with other measures might be studied (Losee, 2000). An earlier study on distributed information retrieval provides methods using Average Search Length as a performance measure that can address issues such as the clustering problems discussed above (Losee & Church, 2004).…”
Section: Discussionmentioning
confidence: 99%
“…Determining the relative performance of two different characteristics used in ordering allows one to make decisions about which types of characteristics should be incorporated into a specific system. While ordering performance measures take many forms (Losee, 2000), e.g. precision, recall, average search length, or any of a number of measures of ordering performance, a linear measure that can be directly interpreted as an occurrence of characteristic X produces performance at the same level as occurrences of characteristic Y can be particularly valuable when making decisions about whether characteristic ¡ or characteristic ¢ should be used, along with their associated costs and benefits.…”
Section: Ordering Performance As a Utility Measurementioning
confidence: 99%
“…Since different evaluation measures evaluate different aspects of retrieval behavior, many evaluation measures have come to proliferate, and these measures have been deeply analyzed and criticized [11,8]. For example, Buckley and Voorhees compare evaluation measures based on their query stability [5], and Loose [10] proposes criteria to determine under which conditions different measures agree or disagree.…”
Section: Introductionmentioning
confidence: 99%