2020
DOI: 10.1109/access.2020.3029100
|View full text |Cite
|
Sign up to set email alerts
|

The Effective Methods for Intrusion Detection With Limited Network Attack Data: Multi-Task Learning and Oversampling

Abstract: Recently, many anomaly intrusion detection algorithms have been developed and applied in network security. These algorithms achieve high detection rate on many classical datasets. However, most of them failed to address two challenges: 1) imbalanced traffic data with limited network attack, 2) multiple data sources that are distributed in different terminals. In detail, those algorithms assume that there are sufficient network traffic data to train their models for intrusion detection. Due to the network attac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(4 citation statements)
references
References 33 publications
0
4
0
Order By: Relevance
“…On the other hand, our solution yields a much higher recall rate than most of the previous works did. We also noted that several authors proposed to tackle the imbalanced data problem, such as [11], [21], [32]- [34]. Unfortunately, none of these methods supports learning on multiple data sources.…”
Section: B Performance Comparison With the State-of-the-art Workmentioning
confidence: 99%
“…On the other hand, our solution yields a much higher recall rate than most of the previous works did. We also noted that several authors proposed to tackle the imbalanced data problem, such as [11], [21], [32]- [34]. Unfortunately, none of these methods supports learning on multiple data sources.…”
Section: B Performance Comparison With the State-of-the-art Workmentioning
confidence: 99%
“…5 shows a schematic of such a workflow. CEF-SsL begins by splitting D in F and L: the former, F, is used exclusively to assess the performance on future data 14 ; the latter, L is used for all remaining 'training' operations, because L can serve as basis to generate L, and then treat the remaining samples as unlabelled, representing U.…”
Section: Stage One: Preparementioning
confidence: 99%
“…Zhou and Zhao [37] proposed the Clustered Multitask Learning (CMTL) approach, which described an arbitrary task with multiple representative tasks to give an accurate representation. Sun et al [38] combined MTL with oversampling for intrusion detection. eir method used MTL to learn relevant information from multiple tasks at the same time and then used the learned information for a single task.…”
Section: Model Algorithmsmentioning
confidence: 99%