2019 International Conference on Advances in Computing and Communication Engineering (ICACCE) 2019
DOI: 10.1109/icacce46606.2019.9079993
|View full text |Cite
|
Sign up to set email alerts
|

Big Data Retrieval using HDFS with LZO Compression

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 21 publications
0
5
0
Order By: Relevance
“…After U 11 and V T 11 are recovered from U 11 and V T 11 , the approximate data matrix A can be reconstructed by replacing S with S , as shown in (6).…”
Section: Icipost-2022mentioning
confidence: 99%
See 1 more Smart Citation
“…After U 11 and V T 11 are recovered from U 11 and V T 11 , the approximate data matrix A can be reconstructed by replacing S with S , as shown in (6).…”
Section: Icipost-2022mentioning
confidence: 99%
“…Lossless compression can recover data without loss of accuracy, it is implemented by adjusting the encoding at the binary level [5]. However, lossless compression sacrifices compression performance for accuracy, which is not very well for big data [6]. Within the accuracy requirement, lossy compression can achieve a better result compared with lossless compression [7].…”
Section: Introductionmentioning
confidence: 99%
“…Hbase supports four different compression algorithms, which can be directly applied on the ColumnFamily. That includes SNAPPY [18], LZO [19], LZ4 [20] and GZ [21] compressions. When creating a table every ColumnFamily is defined separately meaning that some families can have a compression algorithm applied to them and some may not.…”
Section: Column-oriented Data Model Propertiesmentioning
confidence: 99%
“…In addition, we adopt LZO compression to reduce the disk I/O for I/O-intensive tasks. 36 The specific LZO compression principle can be found in Reference 36. The designed algorithm is shown in Algorithm 2.…”
Section: Smosa Task Scheduling Algorithmmentioning
confidence: 99%
“…SMOSA not only optimizes the task scheduling process, but also compresses I/O task data. As the Teragen needs to generate a large amount of disk storage data, the use of LZO compression technology can improve the internal data Shuffle process, 36 and a Shuffle process is the most resource-consuming link in task execution. Therefore, when a large amount of data needs to be generated, the data compression method can reduce the time of data transmission and the disk read time, which effectively shortens the execution time of the task.…”
Section: I/o-intensive Tasksmentioning
confidence: 99%