2009
DOI: 10.14778/1687627.1687675
|View full text |Cite
|
Sign up to set email alerts
|

Turbo-charging estimate convergence in DBO

Abstract: DBO is a database system that utilizes randomized algorithms to give statistically meaningful estimates for the final answer to a multi-table, disk-based query from start to finish during query execution. However, DBO's "time 'til utility" (or "TTU"; that is, the time until DBO can give a useful estimate) can be overly large, particularly in the case that many database tables are joined in a query, or in the case that a join query includes a very selective predicate on one or more of the tables, or when the da… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
33
0

Year Published

2013
2013
2022
2022

Publication Types

Select...
4
3
1

Relationship

2
6

Authors

Journals

citations
Cited by 27 publications
(34 citation statements)
references
References 17 publications
(22 reference statements)
1
33
0
Order By: Relevance
“…This is because Verdict reaches a target error bound much earlier by combining its model with the raw answer of the AQP engine. [6,19,24,36,45,66,87]: Instead of continuously refining approximate answers and reporting them to the user, these engines simply take a time-bound from the user, and then they predict the largest sample size that they can process within the requested time-bound; thus, they minimize error bounds within the allotted time. For these engines, Verdict simply replaces the user's original time bound t 1 with a slightly smaller value t 1 − ǫ before passing it down to the AQP engine, where ǫ is the time needed by Verdict for inferring the improved answer and improved error.…”
Section: Deployment Scenariosmentioning
confidence: 99%
“…This is because Verdict reaches a target error bound much earlier by combining its model with the raw answer of the AQP engine. [6,19,24,36,45,66,87]: Instead of continuously refining approximate answers and reporting them to the user, these engines simply take a time-bound from the user, and then they predict the largest sample size that they can process within the requested time-bound; thus, they minimize error bounds within the allotted time. For these engines, Verdict simply replaces the user's original time bound t 1 with a slightly smaller value t 1 − ǫ before passing it down to the AQP engine, where ǫ is the time needed by Verdict for inferring the improved answer and improved error.…”
Section: Deployment Scenariosmentioning
confidence: 99%
“…The database online aggregation literature has its origins in the seminal paper by Hellerstein et al [21]. We can broadly categorize this body of work into system design [30,7,13,2], online join algorithms [20,8,34], and methods to derive confidence bounds [19]. All of this work is targeted at single-node centralized environments.…”
Section: Related Workmentioning
confidence: 99%
“…In particular, sampling has served as one of the most common and generic approaches to approximation of analytical queries [7,8,13,14,15,21,23,30,31]. The simplest form of sampling is simple random sampling with plug-in estimation.…”
Section: Approximate Query Processing (Aqp)mentioning
confidence: 99%
“…Nearly three decades ago, Olken and Rotem [27] introduced random sampling in relational databases as a means to return approximate answers and reduce query response times. A large body of work has subsequently proposed different sampling techniques [7,8,13,14,15,21,23,30,31,36]. All of this Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation