2021
DOI: 10.12694/scpe.v22i4.1945
|View full text |Cite
|
Sign up to set email alerts
|

Performance-efficient Recommendation and Prediction Service for Big Data frameworks focusing on Data Compression and In-memory Data Storage Indicators

Abstract: The MapReduce framework manages Big Data sets by splitting the large datasets into a set of distributed blocks and processes them in parallel. Data compression and in-memory file systems are widely used methods in Big Data processing to reduce resource-intensive I/O operations and improve I/O rate correspondingly. The article presents a performance-efficient modular and configurable decision-making robust service relying on data compression and in-memory data storage indicators. The service consists of Recomme… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 20 publications
0
7
0
Order By: Relevance
“…Building upon these findings, our future endeavors will focus on extending the available comprehensive scalable platform with new air pollution services [12]. This platform will integrate data from various sources, including archives of remote-sensing images and in-situ measurements.…”
Section: Discussionmentioning
confidence: 99%
“…Building upon these findings, our future endeavors will focus on extending the available comprehensive scalable platform with new air pollution services [12]. This platform will integrate data from various sources, including archives of remote-sensing images and in-situ measurements.…”
Section: Discussionmentioning
confidence: 99%
“…In pursuit of pinpointing the most suitable data compression method tailored to specific use cases, the findings of an insightful study (Astsatryan et al, 2021) recommend a comprehensive assessment methodology. This assessment encompasses the estimation of data processing execution times, a meticulous consideration of an array of data compression techniques, and the diverse spectrum of distributed computing clusters characterized by varying node counts and resource allocations.…”
Section: Storage Optimizationmentioning
confidence: 99%
“…The usage of Big Data frameworks, such as Apache Hadoop or Spark, is also quite promising and widely implemented, as the frameworks can provide efficient and distributed environments for big data processing [17]. The paper [18] shows the effectiveness of an adaptive Spark-based remote sensing data on-demand processing method on the cloud, with the data storage Hadoop Distributed File System (HDFS), which is more efficient in means of execution time than Hadoop.…”
Section: Motivationmentioning
confidence: 99%