2018
DOI: 10.48550/arxiv.1809.04198
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Optimization with Non-Differentiable Constraints with Applications to Fairness, Recall, Churn, and Other Goals

Abstract: We show that many machine learning goals, such as improved fairness metrics, can be expressed as constraints on the model's predictions, which we call rate constraints. We study the problem of training non-convex models subject to these rate constraints (or any non-convex and non-differentiable constraints). In the non-convex setting, the standard approach of Lagrange multipliers may fail. Furthermore, if the constraints are non-differentiable, then one cannot optimize the Lagrangian with gradient-based method… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
17
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
3
1

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(17 citation statements)
references
References 21 publications
0
17
0
Order By: Relevance
“…Interestingly, this reparameterization naturally leads to a Lagrangian based procedure where existing SGD based methods can be employed with little to no change, see Figure 2. While recent results suggest that optimizing nondecomposable data-dependent regularizers may be challenging Cotter et al (2018), our development shows that a sizable subclass of such regularizers indeed admit simple solution schemes. Our overall procedure comes with convergence rate guarantees and optimal per-iteration complexity.…”
Section: Introductionmentioning
confidence: 80%
See 2 more Smart Citations
“…Interestingly, this reparameterization naturally leads to a Lagrangian based procedure where existing SGD based methods can be employed with little to no change, see Figure 2. While recent results suggest that optimizing nondecomposable data-dependent regularizers may be challenging Cotter et al (2018), our development shows that a sizable subclass of such regularizers indeed admit simple solution schemes. Our overall procedure comes with convergence rate guarantees and optimal per-iteration complexity.…”
Section: Introductionmentioning
confidence: 80%
“…Moreover, Column 2 compares Alg. 3 with classical dual ascent procedures from Cotter et al (2018). Here, full projections refers to computing Π C (•) after every inner iteration in Alg.…”
Section: Experimental Evaluationsmentioning
confidence: 99%
See 1 more Smart Citation
“…In Goh et al [14] and Zafar et al [33,34], suitable majorization-minimization/convexconcave procedures [24] were derived. Furthermore, such constrained optimization approaches may lead to more unstable training, and often yield classifiers with both worse accuracy and more unfair [7]. Our proposed method can be solved by off-the-shelf packages, for example, we can use GPC packages by Dezfouli and Bonilla [8] or Gardner et al [13], which only need conditional likelihood evaluation as a black-box function.…”
Section: Introductionmentioning
confidence: 99%
“…Other than post-processing, constrained optimization provides another common approach, in which enforcing group parity is framed as a constraint on a minimum loss objective. The resulting constrained optimization problem is then solved through the use of Lagrange multipliers [22,12,9,5,6,1,16].…”
Section: Introductionmentioning
confidence: 99%