2012 Proceedings IEEE INFOCOM 2012
DOI: 10.1109/infcom.2012.6195649
|View full text |Cite
|
Sign up to set email alerts
|

SubFlow: Towards practical flow-level traffic classification

Abstract: Many research efforts propose the use of flowlevel features (e.g., packet sizes and inter-arrival times) and machine learning algorithms to solve the traffic classification problem. However, these statistical methods have not made the anticipated impact in the real world. We attribute this to two main reasons: (a) training the classifiers and bootstrapping the system is cumbersome, (b) the resulting classifiers have limited ability to adapt gracefully as the traffic behavior changes. In this paper, we propose … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
20
0

Year Published

2012
2012
2021
2021

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 31 publications
(20 citation statements)
references
References 19 publications
0
20
0
Order By: Relevance
“…Since then many papers have attempted to analyse traffic types and perform coarse (identifying traffic type) and/or fine (also identifying the application used) traffic classification using a variety of techniques: a) Nearest Neighbour: Where a decision is made on a testing point's assignment based on its proximity to neighbouring points in the training data [14], [15]. b) Clustering approaches: k-means [16], Density-based spatial clustering (DB-SCAN) [16], subspace clustering [17]. c) Discriminant analysis: Linear Discriminant Analysis (LDA) [14], Quadratic Discriminant Analysis (QDA) [14], Support Vector Machines [18].…”
Section: Related Workmentioning
confidence: 99%
“…Since then many papers have attempted to analyse traffic types and perform coarse (identifying traffic type) and/or fine (also identifying the application used) traffic classification using a variety of techniques: a) Nearest Neighbour: Where a decision is made on a testing point's assignment based on its proximity to neighbouring points in the training data [14], [15]. b) Clustering approaches: k-means [16], Density-based spatial clustering (DB-SCAN) [16], subspace clustering [17]. c) Discriminant analysis: Linear Discriminant Analysis (LDA) [14], Quadratic Discriminant Analysis (QDA) [14], Support Vector Machines [18].…”
Section: Related Workmentioning
confidence: 99%
“…With the increasing use of encryption and the emphasis on privacy, the latter approach using flow statistics with no need for inspecting payloads has been widely studied, with supervised learning with well-known classifiers [12] and semi-supervised learning based on clustering [28,10,4]. While supervised learning is known to yield greater accuracy, supplying labelled data is not a trivial requirement, which made clustering to be spot-lighted.…”
Section: Related Workmentioning
confidence: 99%
“…They used an extended MOGA in feature selection and cluster count optimization for K-Means, resulting in that the detection rate got an increase of 2% to 5%, while the FPR did not increased significantly. Xie G et.al [26] used subspace clustering to make the new classifier learn to identify each application separately just using its own relevant features instead of distinguishing one application from another using the unified feature sets. The approach showed very high accuracy on five traces from different ISPs, and was adaptable to change.…”
Section: Accurate Classification Explorationsmentioning
confidence: 99%