2015 11th International Conference on Computational Intelligence and Security (CIS) 2015
DOI: 10.1109/cis.2015.27
|View full text |Cite
|
Sign up to set email alerts
|

Network Traffic Classification with Improved Random Forest

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 23 publications
(7 citation statements)
references
References 18 publications
0
7
0
Order By: Relevance
“…Input: Imbalanced train set S, scaling factor K, instance hardness threshold IH′, and sample threshold UB Output: New train set S N (1) Step1: Distinguish between easy sets and difcult sets for each sample∈ S do (2) Compute its K nearest neighbors and IH if IH > IH′ then (3) Put the samples into the difcult set (4) end (5) end (6) Difcult set S D and easy set S E � S − S D (7) Step2: Compress the majority samples in the difcult set by the cluster centroid (8) Take all the majority samples from S D and set it as S Maj (9) Use the K-means algorithm with K cluster (10) Use the coordinates of K cluster centroids and replace the majority samples in S Maj (11) Compressed the majority sample set S Maj (12) Step3: Sample the minority samples in the difcult set using the SMOTE algorithm (13) Take all the majority samples from S D and set it as S Min (14) for each sample ∈ S Min do (15) Using SMOTE sampling, the sampling threshold is set to UB (16) Putting new samples into S Z (17) end (18) Step : Merge sample sets (19) Precision is the ratio of the number of samples with positive real values to the number of samples predicted to be positive, which can represent the ability of the model to predict positive samples as follows:…”
Section: Evaluation Metrics and Baseline Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Input: Imbalanced train set S, scaling factor K, instance hardness threshold IH′, and sample threshold UB Output: New train set S N (1) Step1: Distinguish between easy sets and difcult sets for each sample∈ S do (2) Compute its K nearest neighbors and IH if IH > IH′ then (3) Put the samples into the difcult set (4) end (5) end (6) Difcult set S D and easy set S E � S − S D (7) Step2: Compress the majority samples in the difcult set by the cluster centroid (8) Take all the majority samples from S D and set it as S Maj (9) Use the K-means algorithm with K cluster (10) Use the coordinates of K cluster centroids and replace the majority samples in S Maj (11) Compressed the majority sample set S Maj (12) Step3: Sample the minority samples in the difcult set using the SMOTE algorithm (13) Take all the majority samples from S D and set it as S Min (14) for each sample ∈ S Min do (15) Using SMOTE sampling, the sampling threshold is set to UB (16) Putting new samples into S Z (17) end (18) Step : Merge sample sets (19) Precision is the ratio of the number of samples with positive real values to the number of samples predicted to be positive, which can represent the ability of the model to predict positive samples as follows:…”
Section: Evaluation Metrics and Baseline Methodsmentioning
confidence: 99%
“…Te SD sampling algorithm combines oversampling and undersampling methods and considers the spatial distribution of samples during sampling, which overcomes the overgeneralization problem of the SMOTE algorithm to some extent. (2) We propose a two-layer structure combined with XGBoost [10] and the random forest [11] to realize multiclassifcation of trafc, which improves the detection rate and generalization ability of the model. (3) We evaluate the performance of the SD sampling algorithm and the proposed classifcation model using the CICIDS2017 dataset [12].…”
Section: Key Contributions and Papermentioning
confidence: 99%
“…In order to verify the advantages of the 1DCAE-IndRNN method proposed in this paper for malware traffic detection, two classical machine learning methods and three deep learning methods are selected from the literature for experimental comparison. Classical machine learning methods include random forest (RF) [16] and XGBoost [17]. Deep learning methods include deep neural networks (DNN) [18], recurrent neural networks (RNN) [19], and long short-term memory (LSTM) [20].…”
Section: Comparison With Other Methodsmentioning
confidence: 99%
“…In [14], a Naive Bayes technique was used for cell classification based on their network traffic patterns. Also, a random forest-based approach was used for traffic pattern detection in the application layer [15]. Finally, in [4], the authors proposed an approach that combined unsupervised and supervised techniques for the analysis of the performance of a mobile network by monitoring the real-time traffic to detect possible changes.…”
Section: Related Workmentioning
confidence: 99%