2021
DOI: 10.1145/3488269
|View full text |Cite
|
Sign up to set email alerts
|

Towards a Consistent Interpretation of AIOps Models

Abstract: Artificial Intelligence for IT Operations (AIOps) has been adopted in organizations in various tasks, including interpreting models to identify indicators of service failures. To avoid misleading practitioners, AIOps model interpretations should be consistent (i.e., different AIOps models on the same task agree with one another on feature importance). However, many AIOps studies violate established practices in the machine learning community when deriving interpretations, such as interpreting models with subop… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(13 citation statements)
references
References 58 publications
0
13
0
Order By: Relevance
“…We choose a random forest model as it provides an optimal balance between predictive power and explainability. Such models are equipped to fit data better than simple regression models, potentially leading to more robust and reliable insights (Lyu et al 2021 ). Although they are not inherently explainable, several feature attribution methods can and have been used to derive explanations using these models (Breiman 2001 ; Lundberg et al 2018 ; Tantithamthavorn and Jiarpakdee 2021 ).…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…We choose a random forest model as it provides an optimal balance between predictive power and explainability. Such models are equipped to fit data better than simple regression models, potentially leading to more robust and reliable insights (Lyu et al 2021 ). Although they are not inherently explainable, several feature attribution methods can and have been used to derive explanations using these models (Breiman 2001 ; Lundberg et al 2018 ; Tantithamthavorn and Jiarpakdee 2021 ).…”
Section: Methodsmentioning
confidence: 99%
“…We thus define the training set as the first 28 days of data, the validation set as the 29 th day, and the test set as the final day. We split the data along its temporal nature to avoid data leakage (Lyu et al 2021 ). We then use the training set to generate 100 bootstrap samples (i.e., samples with replacement and of the same length as the training set).…”
Section: Methodsmentioning
confidence: 99%
“…We choose a random forest model as it provides an optimal balance between predictive power and explainability. Such models are equipped to fit data better than simple regression models, potentially leading to more robust and reliable insights [25]. Although they are not inherently explainable, several feature attribution methods can and have been used to derive explanations using these models [9,24,46].…”
Section: Closest_tx_pr_time•mentioning
confidence: 99%
“…We thus define the training set as the first 28 days of data, the validation set as the 29 th day, and the test set as the final day. We split the data along its temporal nature to avoid data leakage [25]. We then use the training set to generate 100 bootstrap samples (i.e., samples with replacement and of the same length as the training set).…”
Section: Model Validationmentioning
confidence: 99%
“…One solution that ML practitioners adopt to handle the evolving character of data is retraining/updating ML models over time [24]. Periodical model retraining has also been studied for failure detection AIOps solutions [22], [21] and has proved that continuous model updates achieve better performance over time compared to non-updated models. However, the effects of continuously updated models have only been studied for failure prediction models [21], [22].…”
Section: Introductionmentioning
confidence: 99%