2003
DOI: 10.1007/978-3-540-45248-5_13
|View full text |Cite
|
Sign up to set email alerts
|

An Analysis of the 1999 DARPA/Lincoln Laboratory Evaluation Data for Network Anomaly Detection

Abstract: Abstract. The DARPA/MIT Lincoln Laboratory off-line intrusion detection evaluation data set is the most widely used public benchmark for testing intrusion detection systems. Our investigation of the 1999 background network traffic suggests the presence of simulation artifacts that would lead to overoptimistic evaluation of network anomaly detection systems. The effect can be mitigated without knowledge of specific artifacts by mixing real traffic into the simulation, although the method requires that both the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
191
0
3

Year Published

2007
2007
2021
2021

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 369 publications
(201 citation statements)
references
References 13 publications
0
191
0
3
Order By: Relevance
“…McHugh [McH01] published a strong criticism on the procedures used when creating the KDD'99 database, especially on the lack of verification of the network realism compared to an actual one. It was followed in 2003 by Mahoney et Chan [MC03] who decided to review into detail the database. Mahoney et Chan showed that the traces were far from simulating realistic conditions, and therefore, that even a very simple IDS can exhibit very high performance results, performances that it could never reach in a real environment.…”
Section: I2 the Kdd'99 Traditional Evaluation Methodsmentioning
confidence: 99%
“…McHugh [McH01] published a strong criticism on the procedures used when creating the KDD'99 database, especially on the lack of verification of the network realism compared to an actual one. It was followed in 2003 by Mahoney et Chan [MC03] who decided to review into detail the database. Mahoney et Chan showed that the traces were far from simulating realistic conditions, and therefore, that even a very simple IDS can exhibit very high performance results, performances that it could never reach in a real environment.…”
Section: I2 the Kdd'99 Traditional Evaluation Methodsmentioning
confidence: 99%
“…The DARPA corpus provides a widely-used benchmark for ID evaluation on network traffic at the packet level. Although some works have raised questions about the accuracy and reliability of this dataset [73,74], the DARPA dataset still is the standard corpus for evaluation of NIDSs.…”
Section: Darpa Datasetmentioning
confidence: 99%
“…Presently the only (large) public data set for testing intrusion detection systems is the DARPA data set [17], dated back to 1999. Although this data set is still widely used (since public data sets are scarce), it presents significant shortcomings that make it unsuitable to test our system: E.g., only four attacks related to web (and most of them target web server's vulnerabilities) are available and traffic typology is outdated (see [18,19] for detailed explanations about its limitations). So, to carry out our experiments we collected three different data sets from three different sources:…”
Section: Benchmarksmentioning
confidence: 99%