2015 IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS) 2015
DOI: 10.1109/ispass.2015.7095818
|View full text |Cite
|
Sign up to set email alerts
|

Characterization and analysis of a web search benchmark

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
4
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
3
1

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 14 publications
0
4
0
Order By: Relevance
“…This work extends and earlier work [9] by using larger index size (10GB) and Poisson inter-arrival distribution for dictating query arrival rate. Furthermore, this work provides a broader top-down benchmark characterization, as it performs, among other, (a) an analysis of the performance benefit from various micro-architectural features such as caching and prefetching, and (b) an analysis of the the latency impact of DVFS and idle states.The specific contributions of this work are the following:(1) We characterize end-to-end query processing times and confirm that index search is the most time-consuming part of the query execution [9,27,30].(2) We show that index search time scales linearly with index size, whereas other operations, such as document summary generation, take constant processing time.(3) We demonstrate that a 10GB index, generated with a typical crawl walk starting from various seed web sites, exhibits good load balancing in terms of the number of indexed documents per partition. (4) Given the performance scaling with dataset partitioning and parallel search (due to the good load balancing of index terms across partitions), we motivate the use of low power servers with many simple cores for index search.…”
mentioning
confidence: 81%
See 1 more Smart Citation
“…This work extends and earlier work [9] by using larger index size (10GB) and Poisson inter-arrival distribution for dictating query arrival rate. Furthermore, this work provides a broader top-down benchmark characterization, as it performs, among other, (a) an analysis of the performance benefit from various micro-architectural features such as caching and prefetching, and (b) an analysis of the the latency impact of DVFS and idle states.The specific contributions of this work are the following:(1) We characterize end-to-end query processing times and confirm that index search is the most time-consuming part of the query execution [9,27,30].(2) We show that index search time scales linearly with index size, whereas other operations, such as document summary generation, take constant processing time.(3) We demonstrate that a 10GB index, generated with a typical crawl walk starting from various seed web sites, exhibits good load balancing in terms of the number of indexed documents per partition. (4) Given the performance scaling with dataset partitioning and parallel search (due to the good load balancing of index terms across partitions), we motivate the use of low power servers with many simple cores for index search.…”
mentioning
confidence: 81%
“…(1) We characterize end-to-end query processing times and confirm that index search is the most time-consuming part of the query execution [9,27,30].…”
mentioning
confidence: 83%
“…On the other hand, some articles investigated benchmarking systems at design-time, i.e., performance metrics that can be calculated before the sourcecode is written. Finally, the type 'analytical and measurement' Metric type Articles Performance (de Souza Pinto et al 2018) (van Eyk et al 2018) (Shukla et al 2017) (Aragon et al 2019) (Bermbach et al 2017) (Bondi 2016) (Martinez-Millana et al 2015) (Tekli et al 2011) (Vasar et al 2012) (Franks et al 2011) (Gesvindr & Buhnova 2019) (Ibrahim et al 2018) (Pandey et al 2017) (Ferreira et al 2016) (Ueda et al 2016) (Hadjilambrou et al 2015) (Amaral et al 2015) (Brummett et al 2015 Table 1 Selected articles in the systematic literature review.…”
Section: Benchmark In Software Engineeringmentioning
confidence: 99%
“…We assume that our application scales with speedup, and we can reduce the total number of processors (while still maintaining the same throughput), by re-partitioning the input data [7].…”
Section: Acknowledgmentsmentioning
confidence: 99%