Proceedings of the 3rd International Workshop on Data Management for End-to-End Machine Learning 2019
DOI: 10.1145/3329486.3329489
|View full text |Cite
|
Sign up to set email alerts
|

Debugging Machine Learning Pipelines

Abstract: Machine learning tasks entail the use of complex computational pipelines to reach quantitative and qualitative conclusions. If some of the activities in a pipeline produce erroneous or uninformative outputs, the pipeline may fail or produce incorrect results. Inferring the root cause of failures and unexpected behavior is challenging, usually requiring much human thought, and is both time consuming and error prone. We propose a new approach that makes use of iteration and provenance to automatically infer the … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
23
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

3
4

Authors

Journals

citations
Cited by 23 publications
(23 citation statements)
references
References 14 publications
(26 reference statements)
0
23
0
Order By: Relevance
“…In what follows, we give a brief overview of our debugging methodology. For a more detailed discussion, see [14,15]. splits it into training and test subsets, creates and executes an estimator, and computes the F-measure score using 10-fold cross-validation.…”
Section: Bugdocmentioning
confidence: 99%
See 1 more Smart Citation
“…In what follows, we give a brief overview of our debugging methodology. For a more detailed discussion, see [14,15]. splits it into training and test subsets, creates and executes an estimator, and computes the F-measure score using 10-fold cross-validation.…”
Section: Bugdocmentioning
confidence: 99%
“…In previous work [14], we proposed and implemented new methods to debug machine learning pipelines that automatically and iteratively identify one or more minimal causes of failures, thereby avoiding the tedious and error-prone task of manually tuning and executing new pipeline instances to test and derive new hypotheses for the failures. We have extended this initial work and built BugDoc [15], a system that identifies root causes for errors in general computational pipelines (or workflows).…”
Section: Introductionmentioning
confidence: 99%
“…The first, called Shortcut, discovers definitive root causes (which we sometimes abbreviate to, simply, bugs) consisting of a single conjunction of parameter-value (formally, parameter-equality-value) pairs. The second, called Debugging Decision Trees and introduced in [36], discovers more complex definitive root causes involving inequalities (e.g., A takes a value between 5 and 13).…”
Section: Debugging Algorithmsmentioning
confidence: 99%
“…While the Shortcut and Stacked Shortcut algorithms can find a single minimal definitive root cause very efficiently, usually without truncation (as we will see in the experimental section), characterizing all minimal definitive root causes is challenging. For this purpose, we use an algorithm that is exponential (in the number of parameters) in the worst case, but can characterize inequalities as well as equalities and does well heuristically even with a small budget [36].…”
Section: Finding Bugs With Inequalities: Debugging Decision Treesmentioning
confidence: 99%
“…Basically, there are two approaches to make ML workflows provenance aware. The first is provenance provided for a specific ML platform [46,35,24,40,2,20,36,30] and the second is the provenance systems that are independent of the domain [32]. In the first approach, each ML platform provides provenance using its proprietary representation, which is difficult to interpret and compare with execution between different platforms.…”
Section: Introductionmentioning
confidence: 99%