2020
DOI: 10.1155/2020/4365356
|View full text |Cite
|
Sign up to set email alerts
|

A Log-Based Anomaly Detection Method with Efficient Neighbor Searching and Automatic K Neighbor Selection

Abstract: Using the k-nearest neighbor (kNN) algorithm in the supervised learning method to detect anomalies can get more accurate results. However, when using kNN algorithm to detect anomaly, it is inefficient at finding k neighbors from large-scale log data; at the same time, log data are imbalanced in quantity, so it is a challenge to select proper k neighbors for different data distributions. In this paper, we propose a log-based anomaly detection method with efficient selection of neighbors and automatic selection … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 21 publications
0
4
0
Order By: Relevance
“…The logtransformation plays a crucial role here, enabling low concentrations with values of tenths and hundredths of mg/l or µg/l to be separated from each other. This gives the method added value over outlier detection methods that do not use log-transformation (Adikaram et al, 2015;Wang et al, 2020).…”
Section: Discussionmentioning
confidence: 99%
“…The logtransformation plays a crucial role here, enabling low concentrations with values of tenths and hundredths of mg/l or µg/l to be separated from each other. This gives the method added value over outlier detection methods that do not use log-transformation (Adikaram et al, 2015;Wang et al, 2020).…”
Section: Discussionmentioning
confidence: 99%
“…However, its main drawback lies in its dependence on the K parameter, which can significantly affect classification results and its sensitivity to outlier data. So it is necessary to adjust the K parameter to improve optimal accuracy results (Wang et al, 2020). The basic principle of K-nearest Neighbors (KNN) is to find the closest data to the evaluation data based on the K nearest neighbors in the training dataset.…”
Section: Literature Reviewmentioning
confidence: 99%
“…The kNN classifier is based on a distance function that measures the difference or similarity between two instances (Wang et al, 2020). The standard Euclidean distance "d (x, y)" between two instances "x" and "y" is defined by the formula:…”
Section: Knn Algorithmmentioning
confidence: 99%
“…Therefore, for kNN, there is no actual learning phase. This is why it is generally classified as a lazy learning method (Wang et al, 2020).…”
Section: Knn Algorithmmentioning
confidence: 99%