2008
DOI: 10.1016/j.ejor.2007.08.008
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive data reduction for large-scale transaction data

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2008
2008
2021
2021

Publication Types

Select...
7
1

Relationship

2
6

Authors

Journals

citations
Cited by 32 publications
(8 citation statements)
references
References 22 publications
0
8
0
Order By: Relevance
“…This means, in our problem context, as the amount and sensitivity of data increase, the data consumer’s utilities increase, but at a decreasing rate. This observation is well-grounded on the results of numerous prior analytical and empirical studies in the same or a similar context [23,24,9,19]. A typical example is the use of poll to estimate public opinion.…”
Section: The Proposed Pricing Schemementioning
confidence: 80%
“…This means, in our problem context, as the amount and sensitivity of data increase, the data consumer’s utilities increase, but at a decreasing rate. This observation is well-grounded on the results of numerous prior analytical and empirical studies in the same or a similar context [23,24,9,19]. A typical example is the use of poll to estimate public opinion.…”
Section: The Proposed Pricing Schemementioning
confidence: 80%
“…Obviously, releasing a perturbed (or even unperturbed) sample has lower disclosure risk than releasing the complete data set, because less information is released. However, as far as data utility is concerned, data-mining results based on a sample, even unperturbed, could be substantially different from those based on the complete set (Li and Jacob 2005). In terms of methodology, the approach proposed by Gouweleeuw et al (1998) works essentially on individual or blocks of attributes independently, therefore, "the precise effect on more complicated analyses, such as regression models, can be difficult to assess" (Fienberg and McIntyre 2004, p. 24).…”
Section: Related Workmentioning
confidence: 96%
“…In spite of this, there have been few studies focused on instance selection (or data reduction) for text classification. That is, if too many instances (i.e., documents) are adopted, it can result in large memory requirements and slow execution speed, and can cause over-sensitivity to noise [21,30]. Furthermore, one problem with using the original data points is that there may not be any located at the precise points that would make for the most accurate and concise concept description [23].…”
Section: Introductionmentioning
confidence: 99%