2020
DOI: 10.1007/s13222-020-00335-x
|View full text |Cite
|
Sign up to set email alerts
|

Evaluation Infrastructures for Academic Shared Tasks

Abstract: Academic search systems aid users in finding information covering specific topics of scientific interest and have evolved from early catalog-based library systems to modern web-scale systems. However, evaluating the performance of the underlying retrieval approaches remains a challenge. An increasing amount of requirements for producing accurate retrieval results have to be considered, e.g., close integration of the system's users. Due to these requirements, small to mid-size academic search systems cannot eva… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
1
1

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 22 publications
(21 reference statements)
0
4
0
Order By: Relevance
“…Following previous living lab experiments, we implement the interleaving method by the Team-Draft-Interleaving algorithm [12]. More specifically, we refactored exactly the same implementation 13 for the highest degree of comparability. Furthermore we follow Gingstad et al's proposal of a weighted score based on click events [5] and define the Reward as…”
Section: Evaluation Metricsmentioning
confidence: 99%
See 1 more Smart Citation
“…Following previous living lab experiments, we implement the interleaving method by the Team-Draft-Interleaving algorithm [12]. More specifically, we refactored exactly the same implementation 13 for the highest degree of comparability. Furthermore we follow Gingstad et al's proposal of a weighted score based on click events [5] and define the Reward as…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…Progress in the field of academic search and its corresponding domains is usually evaluated by means of shared tasks that are based on the principles of Cranfield/TREC-style studies [13]. Most recently the TREC-COVID evaluation campaign run by NIST attracted a high number of participants and showed the high impact of scientific retrieval tasks in the community.…”
Section: Introductionmentioning
confidence: 99%
“…Through the STELLA infrastructure the experimental systems of both types could be integrated into the live systems. Further, STELLA creates an interleaved ranking by systematically combining the results from two systems [8,22]. More lifelike results and insights are expected by utilizing real user interactions to assess search systems.…”
Section: Related Workmentioning
confidence: 99%
“…The Living Labs for Academic Search (LiLAS) workshop fosters the discussion, research, and evaluation of academic search systems, and it employs the concept of living labs to the domain of academic search [13]. The goal is to expand the knowledge on improving the search for academic resources like literature, research data, and the interlinking between these resources.…”
Section: Introductionmentioning
confidence: 99%