2020
DOI: 10.24251/hicss.2020.031
|View full text |Cite
|
Sign up to set email alerts
|

Dissecting Moneyball: Improving Classification Model Interpretability in Baseball Pitch Prediction

Abstract: Data science, where technical expertise meets domain knowledge, is collaborative by nature. Complex machine learning models have achieved human-level performance in many areas, yet they face adoption challenges in practice due to limited interpretability of model outputs, particularly for users who lack specialized technical knowledge. One key question is how to unpack complex classification models by enhancing their interpretability to facilitate collaboration in data science research and application. In this… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 13 publications
0
2
0
Order By: Relevance
“…to predict different aspects of baseball using sabermetrics. For instance, Lee et al and Hickey et al used ML models to predict a thrown pitch's outcome (2,3). Furthermore, another study by Bock used sabermetrics and ML models to predict pitchers' short-term and long-term efficacy on their particular teams (4).…”
Section: Articlementioning
confidence: 99%
“…to predict different aspects of baseball using sabermetrics. For instance, Lee et al and Hickey et al used ML models to predict a thrown pitch's outcome (2,3). Furthermore, another study by Bock used sabermetrics and ML models to predict pitchers' short-term and long-term efficacy on their particular teams (4).…”
Section: Articlementioning
confidence: 99%
“…Interpretable machine learning techniques can be characterized in three dimensions: model‐specificity (specific vs. agnostic), generalizability/scope (local vs. global) (Rai, 2020), and stage of data generation (ranging from data to results), as shown in Figure 1. Obtaining model‐specific global explanations (see upper‐left quadrant in Figure 1) is the most developed subarea in interpretable machine learning (Hickey et al, 2020). For instance, feature importance scoring is widely employed to understand an individual feature's contribution to the model.…”
Section: The Kta Perspective Of Complex Machine Learning Modelsmentioning
confidence: 99%
“…To generate interpretations at a user-defined level (e.g., a few instances), one solution is to first enumerate the interpretations on all instances, and then aggregate them (Hickey et al, 2020). To improve the scalability of interpretation aggregation, we propose an approach that identifies the most indicative features in the following steps.…”
Section: Instance Aggregationmentioning
confidence: 99%