1997
DOI: 10.1109/69.591455
|View full text |Cite
|
Sign up to set email alerts
|

Block-oriented compression techniques for large statistical databases

Abstract: Abstract-Disk I/O has long been a performance bottleneck for very large databases. Database compression can be used to reduce disk I/O bandwidth requirements for large data transfers. In this paper, we explore the compression of large statistical databases and propose techniques for organizing the compressed data such that standard database operations such as retrievals, inserts, deletes and modifications are supported. We examine the applicability and performance of three methods. Two of these are adaptations… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
1
0

Year Published

2001
2001
2015
2015

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 48 publications
(4 citation statements)
references
References 19 publications
0
1
0
Order By: Relevance
“…Binary packing is closely related to Frame-Of-Reference (FOR) from Goldstein et al [39] and tuple differential coding from Ng and Ravishankar [40]. In such techniques, arrays of values are partitioned into blocks (e.g., of 128 integers).…”
Section: Binary Packingmentioning
confidence: 99%
“…Binary packing is closely related to Frame-Of-Reference (FOR) from Goldstein et al [39] and tuple differential coding from Ng and Ravishankar [40]. In such techniques, arrays of values are partitioned into blocks (e.g., of 128 integers).…”
Section: Binary Packingmentioning
confidence: 99%
“…Other approaches combined the physical design of a database and compression. Ng and Ravishankar proposed to compress data within large statistical databases by storing the difference between tuples rather than storing all the tuples (Ng & Ravishankar, 1997). The authors in (Kimura, Narasayya, & Syamala, 2011) proposed to use compression in order to guide the physical design of a database and to choose the right auxiliary database structures (indices, materialized view, etc.)…”
Section: Related Workmentioning
confidence: 99%
“…The synthetic data sets were modeled on the criteria used in (Ng & Ravishankar, 1997). Two parameters, degree of skew and degree of variation were used in their generation.…”
Section: Comparison Of the Prime Factor And Wavelet Schemesmentioning
confidence: 99%
“…This effectively ruled out standard compression techniques such as Huffman coding (Cormack, 1985), LZW and its variants (Lempel & Ziv, 1977;Hunt 1998) Arithmetic Coding (Langdon, 1984). These schemes enable decoding to the original data with 100% accuracy, but suffer from modest compression ratios (Ng & Ravishankar, 1997). On the other hand, the trends analysis nature of decision making means that query results do not need to reflect 100% accuracy.…”
Section: Introductionmentioning
confidence: 99%