2017
DOI: 10.1186/s40537-017-0107-2
|View full text |Cite
|
Sign up to set email alerts
|

Scaling associative classification for very large datasets

Abstract: IntroductionIn the recent years, Big Data have received much attention by both the academic and the industrial world, with the aim of fully leveraging the power of the information they hide. The dimensions on which very large datasets usually extend are mainly the size, i.e. the disk storage occupied, the volume, i.e. the number of records, the dimensionality, i.e. the number of features a record can have, and the domain, i.e. the number of distinct values a feature can take. A special effort has been dedicate… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 13 publications
(13 citation statements)
references
References 22 publications
0
11
0
Order By: Relevance
“…where Supp(X =>Y) denotes the support of pattern <if X then Y > while P(X ∪ Y) represents the probability of occurrence of itemset X with class label Y (Hadi, Al-Radaideh & Alhawari, 2018;Nguyen et al, 2018). Definition 7 Confidence of a pattern (X => Y) (Venturini, Baralis & Garza, 2018;Hadi, Al-Radaideh & Alhawari, 2018) is calculated as:…”
Section: Basic Terms Of Associative Classificationmentioning
confidence: 99%
“…where Supp(X =>Y) denotes the support of pattern <if X then Y > while P(X ∪ Y) represents the probability of occurrence of itemset X with class label Y (Hadi, Al-Radaideh & Alhawari, 2018;Nguyen et al, 2018). Definition 7 Confidence of a pattern (X => Y) (Venturini, Baralis & Garza, 2018;Hadi, Al-Radaideh & Alhawari, 2018) is calculated as:…”
Section: Basic Terms Of Associative Classificationmentioning
confidence: 99%
“…• DAC [32]. Ensemble learning which distributes the training of an associative classifier among parallel workers.…”
Section: Big Data Algorithmsmentioning
confidence: 99%
“…DM contains a rich set of classification models; specifically, Support Vector Machine [9], Rule Based [10], Decision Tree [11,12], Bayesian classification [2], k -Nearest Neighbor [13], and AC [14]. Among all, AC is relatively new and promising [15,16,17,18,19,20,21,22,23,24] as it combines the best approaches of association rules mining (ARM) and classification. AC is based on ARM where, first, the strongest Class Association Rules (CAR) are discovered from dataset, followed by converting those rules into classifier model.…”
Section: Introductionmentioning
confidence: 99%
“…After the introduction of AC in 1997, numbers of algorithms are developed in this family e.g. CBA [14,25], CMAR [18], CPAR [21], MCAR [19], MAC [26], CMARAA [27], MRAC & MRAC+ [15], DAC [23], CBA-Spark and CPAR-Spark [24] and G3P-ACBD [28]. Almost all consist of three basic steps -a) Association rule generation, b) Classifier building -rule pruning and rule ranking c) Classification of unknown records using the classifier.…”
Section: Introductionmentioning
confidence: 99%