2021
DOI: 10.48550/arxiv.2110.00603
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Algorithm Fairness in AI for Medicine and Healthcare

Abstract: In the current development and deployment of many artificial intelligence (AI) systems in healthcare, algorithm fairness is a challenging problem in delivering equitable care. Recent evaluation of AI models stratified across race sub-populations have revealed enormous inequalities in how patients are diagnosed, given treatments, and billed for healthcare costs. In this perspective article, we summarize the intersectional field of fairness in machine learning through the context of current issues in healthcare,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 162 publications
(199 reference statements)
1
4
0
Order By: Relevance
“…We measure the performance gap in diagnosis AUC between the advantaged and disadvantaged subgroups as an indicator of group fairness. This is in line with the "separability" criteria (Chen et al, 2021;Dwork et al, 2012) that algorithm scores should be conditionally independent of the sensitive attribute given the diagnostic label (i.e., Ŷ ⊥ S|Y ), which is also adopted by (Gardner et al, 2019;Fong et al, 2021). On the other hand, Zietlow et al (2022) find that for high-capacity models in computer vision, this is typically achieved by worsening the performance of the advantaged group rather than improving the disadvantaged group, a phenomenon termed as leveling down in philosophy that has incurred numerous criticisms (Christiano and Braynen, 2008;Brown, 2003;Doran, 2001).…”
Section: Fairness Definition In Medicinesupporting
confidence: 84%
“…We measure the performance gap in diagnosis AUC between the advantaged and disadvantaged subgroups as an indicator of group fairness. This is in line with the "separability" criteria (Chen et al, 2021;Dwork et al, 2012) that algorithm scores should be conditionally independent of the sensitive attribute given the diagnostic label (i.e., Ŷ ⊥ S|Y ), which is also adopted by (Gardner et al, 2019;Fong et al, 2021). On the other hand, Zietlow et al (2022) find that for high-capacity models in computer vision, this is typically achieved by worsening the performance of the advantaged group rather than improving the disadvantaged group, a phenomenon termed as leveling down in philosophy that has incurred numerous criticisms (Christiano and Braynen, 2008;Brown, 2003;Doran, 2001).…”
Section: Fairness Definition In Medicinesupporting
confidence: 84%
“…Apart from interpretability, few studies addressed fairness and robustness issues. In precision medicine, we aim for a fair system that provides personalized and equitable treatment to each individual without any bias [47]. Chen et al [47] gave the example of biased systems in healthcare.…”
Section: Causality In Healthcare Through Scm Frameworkmentioning
confidence: 99%
“…In precision medicine, we aim for a fair system that provides personalized and equitable treatment to each individual without any bias [47]. Chen et al [47] gave the example of biased systems in healthcare. An algorithm trained only on USA cancer pathology data may lead to wrong classification, when deployed on data from Turkish cancer patients, due to protocol variations or population shifts (imbalanced data).…”
Section: Causality In Healthcare Through Scm Frameworkmentioning
confidence: 99%
“…Artificial intelligence (AI), machine learning (ML), and data-driven technologies are expected to deliver novel ways of understanding and improving mental healthcare. 1 In healthcare applications of AI/ML generally, there has been increased focus on the potential for unintended harm arising from biases present in data 2 and resulting from model assumptions. Two striking examples being racial biases in an algorithm deployed to identify increased healthcare needs 3 and commonly used models for estimating renal function (employing standard biostatistical methods) have been shown to be poorly calibrated for estimating kidney disease in people of colour.…”
Section: Introductionmentioning
confidence: 99%