论文标题

羊乳酪:公平性验证,培训和预测神经网络的算法

FETA: Fairness Enforced Verifying, Training, and Predicting Algorithms for Neural Networks

论文作者

Mohammadi, Kiarash, Sivaraman, Aishwarya, Farnadi, Golnoosh

论文摘要

由神经网络驱动的算法决策在直接影响人们生活质量的应用中变得非常突出。在本文中,我们研究了验证,培训并保证神经网络模型的个人公平性的问题。实施公平性的一种流行方法是将公平概念转化为对模型参数的约束。但是,这种翻译并不总是保证对训练有素的神经网络模型的公平预测。为了应对这一挑战,我们开发了一种反例引导的后处理技术,以在预测时间实现公平限制。与仅在测试或火车数据周围的点上实现公平性的先前工作相反,我们能够在输入域中的所有点上执行并确保公平性。此外,我们提出了一种内部处理技术,将公平性用作归纳偏见,通过迭代地将公平性反例纳入学习过程。我们已经在称为Feta的工具中实施了这些技术。对现实世界数据集的经验评估表明,羊乳酪不仅能够在预测时间内保持公平性,而且能够训练表现出更高程度的个人公平程度的准确模型。

Algorithmic decision making driven by neural networks has become very prominent in applications that directly affect people's quality of life. In this paper, we study the problem of verifying, training, and guaranteeing individual fairness of neural network models. A popular approach for enforcing fairness is to translate a fairness notion into constraints over the parameters of the model. However, such a translation does not always guarantee fair predictions of the trained neural network model. To address this challenge, we develop a counterexample-guided post-processing technique to provably enforce fairness constraints at prediction time. Contrary to prior work that enforces fairness only on points around test or train data, we are able to enforce and guarantee fairness on all points in the input domain. Additionally, we propose an in-processing technique to use fairness as an inductive bias by iteratively incorporating fairness counterexamples in the learning process. We have implemented these techniques in a tool called FETA. Empirical evaluation on real-world datasets indicates that FETA is not only able to guarantee fairness on-the-fly at prediction time but also is able to train accurate models exhibiting a much higher degree of individual fairness.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源