Computer Science Faculty Publications and Presentations 2018
DOI: 10.18122/cs_facpubs/148/boisestate
|View full text |Cite
|
Sign up to set email alerts
|

Monte Carlo Estimates of Evaluation Metric Error and Bias

Abstract: Traditional offline evaluations of recommender systems apply metrics from machine learning and information retrieval in settings where their underlying assumptions no longer hold. This results in significant error and bias in measures of top-N recommendation performance, such as precision, recall, and nDCG. Several of the specific causes of these errors, including popularity bias and misclassified decoy items, are well-explored in the existing literature. In this paper we survey a range of work on identifying … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 10 publications
(17 reference statements)
0
1
0
Order By: Relevance
“…In [9], the authors proposed FaiRecSys, an algorithm that mitigates algorithmic bias by post-processing the recommendation matrix with minimum impact on the accuracy of recommendations provided to the end-users. Third, evaluation metric error and bias [33] simulates the recommender data generation and evaluation processes to quantify how erroneous current evaluation practices are. In [35], the authors proposed a simulation framework for measuring the impact of a recommender system under different types of user behavior.…”
Section: Bias Mitigationmentioning
confidence: 99%
“…In [9], the authors proposed FaiRecSys, an algorithm that mitigates algorithmic bias by post-processing the recommendation matrix with minimum impact on the accuracy of recommendations provided to the end-users. Third, evaluation metric error and bias [33] simulates the recommender data generation and evaluation processes to quantify how erroneous current evaluation practices are. In [35], the authors proposed a simulation framework for measuring the impact of a recommender system under different types of user behavior.…”
Section: Bias Mitigationmentioning
confidence: 99%