Proceedings of the International Workshop on Reproducibility and Replication in Recommender Systems Evaluation 2013
DOI: 10.1145/2532508.2532511
|View full text |Cite
|
Sign up to set email alerts
|

A comparative analysis of offline and online evaluations and discussion of research paper recommender system evaluation

Abstract: Offline evaluations are the most common evaluation method for research paper recommender systems. However, no thorough discussion on the appropriateness of offline evaluations has taken place, despite some voiced criticism. We conducted a study in which we evaluated various recommendation approaches with both offline and online evaluations. We found that results of offline and online evaluations often contradict each other. We discuss this finding in detail and conclude that offline evaluations may be inapprop… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
81
0
1

Year Published

2013
2013
2024
2024

Publication Types

Select...
5
3
1

Relationship

2
7

Authors

Journals

citations
Cited by 106 publications
(82 citation statements)
references
References 25 publications
0
81
0
1
Order By: Relevance
“…The majority (71 %) of approaches were evaluated using offline evaluations, which are subject to various shortcomings. Some claim that offline evaluations should not be used for evaluating research-paper recommender systems [191]. If that is true, most of the reviewed evaluations would be of little significance.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…The majority (71 %) of approaches were evaluated using offline evaluations, which are subject to various shortcomings. Some claim that offline evaluations should not be used for evaluating research-paper recommender systems [191]. If that is true, most of the reviewed evaluations would be of little significance.…”
Section: Discussionmentioning
confidence: 99%
“…3). [ [5][6][7][8]20,25,69,70,112,113,125,159,[189][190][191][192][193][194] We then present the survey of the 96 approaches' evaluations (Sect. 4), followed by an analysis of the 62 approaches that we chose for the in-depth review (Sect.…”
Section: Paper | Article | Citation] [Recommender | Recommendation] [mentioning
confidence: 99%
“…Typically, offline evaluations are considered suitable to preselect a set of promising algorithms, which are subsequently evaluated in online evaluations or by a user study [11]. However, there is serious criticism of offline evaluations [60][61][62][63][64][65]106,111].…”
Section: Evaluating Methodsmentioning
confidence: 99%
“…Filtering these information sources result in the higher accuracy of recommendations. In [8], authors have discussed the online and offline evaluation of research paper recommender framework and conclude that offline evaluation in this domain does not provide promising results. Docear"s research paper recommender framework is proposed using content-based filtering in which user"s data (citations, references, and papers) is directed in mind maps and are then utilized for recommendations [9].…”
Section: Literature Reviewmentioning
confidence: 99%