2020
DOI: 10.1007/978-3-030-54994-7_15
|View full text |Cite
|
Sign up to set email alerts
|

Anomaly Detection from Log Files Using Unsupervised Deep Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 19 publications
(17 citation statements)
references
References 8 publications
0
15
0
Order By: Relevance
“…The most common loss function in the reviewed publications is the Cross-Entropy (CE), in particular, the categorical cross-entropy for multi-class prediction [20], [57] or binary cross-entropy that only differentiates between the normal and anomalous class [61]. Other common loss functions include the Hyper-Sphere Objective Function (HS) where the distance to the center of a hyper-sphere represents the anomaly score [24], [39], [41], [62], the Mean Squared Error (MSE) that is used for regression [20], [27], [28], [47], [50], [53], [68], and the Kullback-Leibler Divergence (KL) and Marginal Likelihood (ML) that are useful to measure loss in probability distributions [49], [58].…”
Section: B Deep Learning Techniquesmentioning
confidence: 99%
See 1 more Smart Citation
“…The most common loss function in the reviewed publications is the Cross-Entropy (CE), in particular, the categorical cross-entropy for multi-class prediction [20], [57] or binary cross-entropy that only differentiates between the normal and anomalous class [61]. Other common loss functions include the Hyper-Sphere Objective Function (HS) where the distance to the center of a hyper-sphere represents the anomaly score [24], [39], [41], [62], the Mean Squared Error (MSE) that is used for regression [20], [27], [28], [47], [50], [53], [68], and the Kullback-Leibler Divergence (KL) and Marginal Likelihood (ML) that are useful to measure loss in probability distributions [49], [58].…”
Section: B Deep Learning Techniquesmentioning
confidence: 99%
“…The time stamp of log events is a special parameter as it allows to put other parameters in temporal context, which is required for time-series analysis. However, the time stamps themselves may be used for Time Embedding (TE) and serve as input to neural networks [27]. For this purpose, Li et al [46] generate vectors for sequences of time differences between event occurrences by applying soft one-hot encoding.…”
Section: Log Data Preparationmentioning
confidence: 99%
“…Table 2: Description of Fields for Syslogs field example description timestamp Jan 1 00:08:35 the time stamp host i151-306 the node where the job ran system-id kernel (linux) ID of the system application Lustre application name text message an error occurred .... 6 detailed message of the event 5.1.2 Rationalized logs. Rationalized logs was a new logging framework for TACC supercomputers instead of Syslogs.…”
Section: Evaluation System and Datasetsmentioning
confidence: 99%
“…Although log files are nontrivial for analysis (e.g., they are often unstructured, duplicated or even incomplete [10]), extensive research on failure-related analysis using HPC system logs has been undertaken such as detecting anomalies e.g., [4], [43], [6], diagnosing the root causes of failures, e.g., [10,14,17], and detecting the errors that lead to system failures, e.g., [3,42,57,81].…”
Section: Introductionmentioning
confidence: 99%
“…In [2], the authors performed error detection in supercomputers by combining entropy, mutual information, and PCA approaches. In general, techniques have focused on capturing anomalies in system logs, e.g., these recent works were based on anomaly detection techniques [3]- [5]). Recently, techniques based on natural language processing (NLP) and artificial intelligence (AI) have been applied towards failure log analysis of these systems [6]- [8].…”
Section: Introductionmentioning
confidence: 99%