Proceedings of the 12th International Conference on Extending Database Technology: Advances in Database Technology 2009
DOI: 10.1145/1516360.1516487
|View full text |Cite
|
Sign up to set email alerts
|

Estimating aggregates in time-constrained approximate queries in Oracle

Abstract: The concept of time-constrained SQL queries was introduced to address the problem of long-running SQL queries. A key approach adopted for supporting time-constrained SQL queries is to use sampling to reduce the amount of data that needs to be processed, thereby allowing completion of the query in the specified time constraint. However, sampling does make the query results approximate and hence requires the system to estimate the values of the expressions (especially aggregates) occurring in the select list. Th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
13
0

Year Published

2010
2010
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 9 publications
(13 citation statements)
references
References 4 publications
0
13
0
Order By: Relevance
“…Besides a few limited benefits (see Previous Approaches below), the work (both I/O and computation) performed for answering past queries is often wasted afterwards. However, in an approximate query processing context (e.g., [6,19,34,36,66,85]), one might be able to change this paradigm and reuse much of the previous work done by the database system based on the following observation: * This manuscript is an extended report of the work published in ACM SIGMOD conference 2017.…”
Section: Introductionmentioning
confidence: 99%
“…Besides a few limited benefits (see Previous Approaches below), the work (both I/O and computation) performed for answering past queries is often wasted afterwards. However, in an approximate query processing context (e.g., [6,19,34,36,66,85]), one might be able to change this paradigm and reuse much of the previous work done by the database system based on the following observation: * This manuscript is an extended report of the work published in ACM SIGMOD conference 2017.…”
Section: Introductionmentioning
confidence: 99%
“…There has been a large body of research on using sampling to provide quick answers to database queries, on database systems [9,15,16,22,23,24,25,33,44], and data stream systems [12,31]. Approximate aggregate processing has been the focus of many of these works, which study randomized joins [24], optimal sample construction [9,16], sample reusing [44], and sampling plan in a stream setting [12,31].…”
Section: Related Workmentioning
confidence: 99%
“…Approximate aggregate processing has been the focus of many of these works, which study randomized joins [24], optimal sample construction [9,16], sample reusing [44], and sampling plan in a stream setting [12,31]. Most of them use statistical inequalities and the central limit theorem to model the confidence interval or variance of the approximate aggregate answers [9,16,22,23,24,44]. Recently, Pansare et al [33] develop a very sophisticated Bayesian framework to infer the confidence bounds of approximate aggregate answers.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…This situation has brought even more attention to the already-active area of Approximate Query Processing (AQP). As a critical and general approach for coping with massive datasets, sampling is widely used in databases [4,6,9,11,12,13,18], Map-Reduce systems [5,16], and data stream management systems [7,17].…”
Section: Introductionmentioning
confidence: 99%