2006
DOI: 10.1007/11827405_12
|View full text |Cite
|
Sign up to set email alerts
|

DCF: An Efficient Data Stream Clustering Framework for Streaming Applications

Abstract: Abstract. Streaming applications, such as environment monitoring and vehicle location tracking require handling high volumes of continuously arriving data and sudden fluctuations in these volumes while efficiently supporting multidimensional historical queries. The use of the traditional database management systems is inappropriate because they require excessive number of disk I/O in continuously updating massive data streams. In this paper, we propose DCF (Data Stream Clustering Framework), a novel framework … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2007
2007
2017
2017

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 12 publications
(8 reference statements)
0
4
0
Order By: Relevance
“…However, density based algorithms are different from fuzzy clustering algorithms as they try to optimize a different objective function. In [21] a framework for efficiently archiving high volumes of streaming data was proposed, which reduces disk access for storing and retrieving data. They grouped incoming data into clusters and stored them instead of raw data.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…However, density based algorithms are different from fuzzy clustering algorithms as they try to optimize a different objective function. In [21] a framework for efficiently archiving high volumes of streaming data was proposed, which reduces disk access for storing and retrieving data. They grouped incoming data into clusters and stored them instead of raw data.…”
Section: Related Workmentioning
confidence: 99%
“…Clustering streaming data is becoming important due to the availability of large amounts of data recorded from everyday transactions, web click streams, telephone records, web documents, sensors data, environmental data, and network monitoring [7], [13], [17], [20], [19], [21]. The sources of streaming data are increasing as technology for recording events is becoming cheaper.…”
Section: Introductionmentioning
confidence: 99%
“…,X i } according to the memory size. A load monitor [53] ensures that the loading of spatial data fits the main memory.…”
Section: Asc-based Stream Clusteringmentioning
confidence: 99%
“…Synthetic data sets are usually preferred because testing hypotheses like noiserobustness, and scaling for high-dimensionality are easier to perform with synthetic data. Examples of artificially generated data sets are: (i) data generated by varying Gaussian distributions [Aggarwal et al 2003;Wan et al 2008;Dang et al 2009]; (ii) data generated by the IBM synthetic data generator [Ong et al 2004]; (iii) data simulating a taxi location tracking application [Cho et al 2006]; and (iv) data sets formed by arbitrarily-shaped clusters, like those presented in Figure 12.…”
mentioning
confidence: 99%