Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence 2021
DOI: 10.24963/ijcai.2021/75
|View full text |Cite
|
Sign up to set email alerts
|

Addressing the Long-term Impact of ML Decisions via Policy Regret

Abstract: Machine Learning (ML) increasingly informs the allocation of opportunities to individuals and communities in areas such as lending, education, employment, and beyond. Such decisions often impact their subjects' future characteristics and capabilities in an a priori unknown fashion. The decision-maker, therefore, faces exploration-exploitation dilemmas akin to those in multi-armed bandits. Following prior work, we model communities as arms. To capture the long-term effects of ML-based allocation decisions, we … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 19 publications
1
4
0
Order By: Relevance
“…Thus far, our motivation for the tallying bandit setting has been primarily theoretical, to resolve the gap in our understanding of when we can efficiently minimize CPR. Nevertheless, in similar vein to Heidari et al [HKR16], Lindner et al [LHK21] and Awasthi et al [ABGK22], we believe that the tallying bandit is a simple approximation for various practical settings. For instance, in recommender systems the reward associated with an action is rarely static, because the stimulus of recommended content influences user preferences [CLA + 03, SGR16].…”
Section: Tallying Banditssupporting
confidence: 82%
See 4 more Smart Citations
“…Thus far, our motivation for the tallying bandit setting has been primarily theoretical, to resolve the gap in our understanding of when we can efficiently minimize CPR. Nevertheless, in similar vein to Heidari et al [HKR16], Lindner et al [LHK21] and Awasthi et al [ABGK22], we believe that the tallying bandit is a simple approximation for various practical settings. For instance, in recommender systems the reward associated with an action is rarely static, because the stimulus of recommended content influences user preferences [CLA + 03, SGR16].…”
Section: Tallying Banditssupporting
confidence: 82%
“…A natural restriction is to enforce that each g x has special structure. This is precisely the approach taken by works on rotting bandits [HKR16, LCM17, SLC + 19, SMLV20], improving bandits [HKR16], single peaked bandits [LHK21] and congested bandits [ABGK22]. Concretely, these works use base functions {g x } x∈X that have the following special "tallying" structure.…”
Section: Restricting the Adversarymentioning
confidence: 99%
See 3 more Smart Citations