2000
DOI: 10.1145/380995.381017
|View full text |Cite
|
Sign up to set email alerts
|

Mining frequent patterns with counting inference

Abstract: In this paper, we propose the algorithm PASCAL which introduces a novel optimization of the well-known algorithm Apriori. This optimization is based on a new strategy called pattern counting inference that relies on the concept of key patterns. We show that the support of frequent non-key patterns can be inferred from frequent key patterns without accessing the database. Experiments comparing PAS-CAL to the three algorithms Apriori, Close and Max-Miner, show that PASCAL is among the most efficient algorithms f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
138
0
1

Year Published

2002
2002
2013
2013

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 222 publications
(139 citation statements)
references
References 21 publications
0
138
0
1
Order By: Relevance
“…Since the number of frequent itemsets can be huge in dense databases, it is now common to use condensed representations (e.g., free itemsets, closed ones, non derivable itemsets [10]) to save space and time during the frequent itemset mining task and to avoid some redundancy. Since [11], it is common to formalize the fact that many itemsets have the same closure by means of closure equivalence relation. Each CEC contains exactly one maximal itemset (w.r.t.…”
Section: Feature Construction Using Closure Equivalence Classesmentioning
confidence: 99%
See 1 more Smart Citation
“…Since the number of frequent itemsets can be huge in dense databases, it is now common to use condensed representations (e.g., free itemsets, closed ones, non derivable itemsets [10]) to save space and time during the frequent itemset mining task and to avoid some redundancy. Since [11], it is common to formalize the fact that many itemsets have the same closure by means of closure equivalence relation. Each CEC contains exactly one maximal itemset (w.r.t.…”
Section: Feature Construction Using Closure Equivalence Classesmentioning
confidence: 99%
“…One breakthrough into the computational complexity of such mining tasks has been obtained thanks to condensed representations for frequent itemsets, i.e., rather small collections of patterns from which one can infer the frequency of many sets instead of counting for it (see [10] for a survey). In this paper, we consider closure equivalence classes, i.e., frequent closed sets and their generators [11]. Furthermore, when considering the δ-free itemsets with δ > 0 [12,13], we can consider a "near equivalence" perspective and thus, roughly speaking, the concept of almost-closed itemsets.…”
Section: Introductionmentioning
confidence: 99%
“…It computes in a level-wise manner all frequent key sets, and in the same step their closures. Pascal [5] differs from Titanic in that it additionally produces all frequent itemsets. [4] discusses efficient data structures for the algorithms Pascal and Titanic.…”
Section: Algorithms For Computing Frequent Closed / Key Itemsetsmentioning
confidence: 99%
“…Since our rules are complete, this shows that additional gain is in many cases unlikely. PASCAL [3] In their PASCAL-algorithm, Bastide et al use counting inference to avoid counting the support of all candidates. The rule they are using to avoid counting is based on our rule R I (I − {i}).…”
Section: Proofmentioning
confidence: 99%