论文标题

到公平边界及以后:识别,量化和优化公平 - 准确性帕累托前沿

To the Fairness Frontier and Beyond: Identifying, Quantifying, and Optimizing the Fairness-Accuracy Pareto Frontier

论文作者

Little, Camille Olivia, Weylandt, Michael, Allen, Genevera I

论文摘要

当使用机器学习做出高风险社会决策时,算法公平已成为一个重要的考虑因素。然而,改善的公平通常以牺牲模型准确性为代价。尽管已经研究了公平准确性权衡的各个方面,但大多数工作分别报告了各种模型的公平和准确性。如果没有模型不足的指标,这几乎使模型比较几乎是不可能的,该指标反映了两个Desiderata的平衡。我们寻求确定,量化和优化公平准确权衡权衡的经验帕累托前沿。具体而言,我们通过折衷方案(TAF)曲线来识别和概述经验的帕累托前沿;然后,我们开发一个通过TAF曲线下的加权区域来量化此帕累托边界的度量,我们将其称为“公平”区域以下的曲线(FAUC)。 TAF曲线提供了Pareto边境的第一个经验性,模型不足的表征,而FAUC为公正地比较模型家族在公平性和准确性上提供了第一个指标。 TAF曲线和FAUC均可采用所有群体公平定义和准确度措施。接下来,我们问:是否有可能扩展经验的帕累托前沿,从而改善给定拟合模型集合的FAUC?我们通过开发一个新颖的公平模型堆叠框架Fairstacks来回答,该框架解决了凸面程序,以最大程度地提高模型集成的准确性。我们表明,通过Fairstacks进行优化总是扩大经验的帕累托前沿并改善了FAUC。我们还研究了我们提出的方法的其他理论特性。最后,我们通过对几个真正的基准数据集的研究来验证TAF,FAUC和Fairstacks,这表明FairStacks会导致FAUC的重大改进,从而超过了现有的算法公平方法。

Algorithmic fairness has emerged as an important consideration when using machine learning to make high-stakes societal decisions. Yet, improved fairness often comes at the expense of model accuracy. While aspects of the fairness-accuracy tradeoff have been studied, most work reports the fairness and accuracy of various models separately; this makes model comparisons nearly impossible without a model-agnostic metric that reflects the balance of the two desiderata. We seek to identify, quantify, and optimize the empirical Pareto frontier of the fairness-accuracy tradeoff. Specifically, we identify and outline the empirical Pareto frontier through Tradeoff-between-Fairness-and-Accuracy (TAF) Curves; we then develop a metric to quantify this Pareto frontier through the weighted area under the TAF Curve which we term the Fairness-Area-Under-the-Curve (FAUC). TAF Curves provide the first empirical, model-agnostic characterization of the Pareto frontier, while FAUC provides the first metric to impartially compare model families on both fairness and accuracy. Both TAF Curves and FAUC can be employed with all group fairness definitions and accuracy measures. Next, we ask: Is it possible to expand the empirical Pareto frontier and thus improve the FAUC for a given collection of fitted models? We answer affirmately by developing a novel fair model stacking framework, FairStacks, that solves a convex program to maximize the accuracy of model ensemble subject to a score-bias constraint. We show that optimizing with FairStacks always expands the empirical Pareto frontier and improves the FAUC; we additionally study other theoretical properties of our proposed approach. Finally, we empirically validate TAF, FAUC, and FairStacks through studies on several real benchmark data sets, showing that FairStacks leads to major improvements in FAUC that outperform existing algorithmic fairness approaches.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源