Proceedings of the 2009 SIAM International Conference on Data Mining 2009
DOI: 10.1137/1.9781611972795.92
|View full text |Cite
|
Sign up to set email alerts
|

Near-optimal supervised feature selection among frequent subgraphs

Abstract: Graph classification is an increasingly important step in numerous application domains, such as function prediction of molecules and proteins, computerised scene analysis, and anomaly detection in program flows.Among the various approaches proposed in the literature, graph classification based on frequent subgraphs is a popular branch: Graphs are represented as (usually binary) vectors, with components indicating whether a graph contains a particular subgraph that is frequent across the dataset.On large graphs… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
97
0

Year Published

2009
2009
2019
2019

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 98 publications
(97 citation statements)
references
References 36 publications
(46 reference statements)
0
97
0
Order By: Relevance
“…We run the discriminatory pattern sampling algorithm with a minimum support value of 6 (20%). The leap search [34], and CORK [23] algorithms did not run on this dataset. Both failed with a segmentation fault.…”
Section: Sampling Results On Large Graphsmentioning
confidence: 99%
See 2 more Smart Citations
“…We run the discriminatory pattern sampling algorithm with a minimum support value of 6 (20%). The leap search [34], and CORK [23] algorithms did not run on this dataset. Both failed with a segmentation fault.…”
Section: Sampling Results On Large Graphsmentioning
confidence: 99%
“…Though majority of these algorithms consider the summarization of itemset patterns, [34,23] consider graph patterns. In another recent work in the graph domain, Hasan et.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…For instance, CBA [1] first computes all frequent itemsets (with their most frequent class label) and then induces an ordered rule-list classifier by removing redundant itemsets. Several alternative techniques (for instance, [28,30]) define measures of redundancy and ways to select only a limited number of patterns. Constructing a concise pattern set for use in classification can be seen as a form of feature selection.…”
Section: Global Heuristic Two Step Techniquesmentioning
confidence: 99%
“…However, their approach is prone to model overfitting as it does not provide any regularization capability. Very recently, a greedy subgraph feature selection algorithm, named CORK [64] was proposed. It is embedded in the gSpan [96] mining process and it can provide an approximation guaranty with respect to a sub-modular quality criteria.…”
Section: Mining Discriminative Patternsmentioning
confidence: 99%