2010
DOI: 10.1145/1670679.1670680
|View full text |Cite
|
Sign up to set email alerts
|

A survey of online failure prediction methods

Abstract: With the ever-growing complexity and dynamicity of computer systems, proactive fault management is an effective approach to enhancing availability. Online failure prediction is the key to such techniques. In contrast to classical reliability methods, online failure prediction is based on runtime monitoring and a variety of models and methods that use the current state of a system and, frequently, the past experience as well. This survey describes these methods. To capture the wide spectrum of approaches concer… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
245
0
4

Year Published

2010
2010
2017
2017

Publication Types

Select...
6
4

Relationship

0
10

Authors

Journals

citations
Cited by 470 publications
(267 citation statements)
references
References 54 publications
0
245
0
4
Order By: Relevance
“…Contingency table metrics [17] are used to investigate how accurately a decision approach works. An adaptation decision approach can terminate with two possible states: warning or silence (refer to Figure 3).…”
Section: B Evaluation Metricsmentioning
confidence: 99%
“…Contingency table metrics [17] are used to investigate how accurately a decision approach works. An adaptation decision approach can terminate with two possible states: warning or silence (refer to Figure 3).…”
Section: B Evaluation Metricsmentioning
confidence: 99%
“…A similar problem of learning in component based systems, however, has been widely studied in systems recognizing the causes for system errors (see surveys in [5] and [6]). We have already discussed an important work [4], in Sections III and IV, which we also use for evaluating our approach.…”
Section: Related Work In Service Composition Learningmentioning
confidence: 99%
“…The development of automated systems for pattern recognition is associated with the need of big data processing [1][2][3][4][5][6]. Typically, the original samples of data describing the objects or processes under investigation may contain redundant and uninformative information [7][8][9][10][11][12][13].…”
Section: Introductionmentioning
confidence: 99%