Boise State ScholarWorks
DOI: 10.18122/b2vx12
|View full text |Cite
|
Sign up to set email alerts
|

Towards Minimal Necessary Data: The Case for Analyzing Training Data Requirements of Recommender Algorithms

Abstract: This paper states the case for the principle of minimal necessary data: If two recommender algorithms achieve the same effectiveness, the better algorithm is the one that requires less user data. Applying this principle involves carrying out training data requirements analysis, which we argue should be adopted as best practice for the development and evaluation of recommender algorithms. We take the position that responsible recommendation is recommendation that serves the people whose data it uses. To minimiz… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
6
0

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 8 publications
(10 citation statements)
references
References 24 publications
2
6
0
Order By: Relevance
“…The authors suggests a differential data analysis for understanding which data contributes to performance in recommender systems, and propose that less useful data should be discarded based on the analysis. While we fully share their motivation and view that performance saturates with data size (as empirically confirmed in [19]), we like to highlight the post-hoc nature of their analysis. The choice of which particular data should be collected and eventually discarded is made after the data has been analysed.…”
Section: Background and Related Workmentioning
confidence: 76%
See 2 more Smart Citations
“…The authors suggests a differential data analysis for understanding which data contributes to performance in recommender systems, and propose that less useful data should be discarded based on the analysis. While we fully share their motivation and view that performance saturates with data size (as empirically confirmed in [19]), we like to highlight the post-hoc nature of their analysis. The choice of which particular data should be collected and eventually discarded is made after the data has been analysed.…”
Section: Background and Related Workmentioning
confidence: 76%
“…We see this for all studied markets. This is on a par with the saturation effect reported in [19]. However, while they conclude a decline of accuracy with the squared error metric on training data, we look at validation performance, with metrics purposely designed for measuring quality of recommendations.…”
Section: Performance Of Size Of Training Datamentioning
confidence: 87%
See 1 more Smart Citation
“…Here, we zero in specifically on data minimization for recommender systems. In [23], the authors proposed to adopt training data requirements analysis to analyze and evaluate the trade-off between the amount of data that the system requires, and the performance of the system. In [21], the authors proposed to extend the data minimzation principles advocated in GDPR and studied their effect on recommender systems.…”
Section: Data Minimizationmentioning
confidence: 99%
“…Recently, the area of recommender systems has seen growing interest in data minimization. For example, in [51] data requirements analysis is proposed as best practice. The argument this work advances is straightforward: just as we avoid developing algorithms with unnecessary computational complexity, we should also avoid developing algorithms that need unnecessary data.…”
Section: Datamentioning
confidence: 99%