Equity and Access in Algorithms, Mechanisms, and Optimization 2021
DOI: 10.1145/3465416.3483305
|View full text |Cite
|
Sign up to set email alerts
|

A Framework for Understanding Sources of Harm throughout the Machine Learning Life Cycle

Abstract: As machine learning (ML) increasingly affects people and society, awareness of its potential unwanted consequences has also grown. To anticipate, prevent, and mitigate undesirable downstream consequences, it is critical that we understand when and how harm might be introduced throughout the ML life cycle. In this paper, we provide a framework that identifies seven distinct potential sources of downstream harm in machine learning, spanning data collection, development, and deployment. In doing so, we aim to fac… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
218
0
6

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 251 publications
(224 citation statements)
references
References 32 publications
(21 reference statements)
0
218
0
6
Order By: Relevance
“…We draw on Suresh and Guttag's [49] framework for understanding sources of harm through the machine learning life cycle to ground our investigation into bias in automated speaker recognition. Suresh and Guttag divide the machine learning life cycle into two streams and identify seven sources of bias related harms across the two streams: 1) the data generation stream can contain historical, representational and measurement bias; and 2) the model building and implementation stream can contain learning, aggregation, evaluation and deployment bias.…”
Section: Framework For Understanding Sources Of Harm Through the Mach...mentioning
confidence: 99%
See 2 more Smart Citations
“…We draw on Suresh and Guttag's [49] framework for understanding sources of harm through the machine learning life cycle to ground our investigation into bias in automated speaker recognition. Suresh and Guttag divide the machine learning life cycle into two streams and identify seven sources of bias related harms across the two streams: 1) the data generation stream can contain historical, representational and measurement bias; and 2) the model building and implementation stream can contain learning, aggregation, evaluation and deployment bias.…”
Section: Framework For Understanding Sources Of Harm Through the Mach...mentioning
confidence: 99%
“…Representative benchmark datasets are particularly important during machine learning development, as benchmarks have disproportionate power to scale bias across applications if models overfit to the data in the benchmark [49]. Three evaluation sets can be constructed from the VoxCeleb 1 dataset to benchmark speaker verification models.…”
Section: Evaluation Biasmentioning
confidence: 99%
See 1 more Smart Citation
“…Unmitigated, these biases transfer into algorithmic decisions that are fed back into the decision making process, thus aggravating the situation. Furthermore, algorithmic decisions themselves can also encode new biases that were not initially present in historical data [72]; this may be due to poor quality [73], and even the careless use of fairness constraints. [74] Like integrity and privacy, the challenges of algorithmic fairness can be considered through the lens of attacks and defenses.…”
Section: Risksmentioning
confidence: 99%
“…For example, in pre-trials where a judge has to assess the risk of accepting or rejecting bail, the defendants are only being subject to algorithmic predictions because they have been arrested in the first place. Mitchell et al [75] provide a broad overview on various fairness risks associated with three broad categories: (i) data bias: both statistical-which evolves from lack of representation (i.e., selective labels [76]), and societal-which indicates a normative mismatch [73]; (ii) model bias: influence of choice of hypothesis [77] and even interpretability, and (iii) evaluation: how today's predictions can alter decision subjects' behaviors, thus requiring new predictions in the future. We will discuss such performative qualities of models in § III-D1.…”
Section: Risksmentioning
confidence: 99%