论文标题

通过校正向量进行公平解释的学习

Fair Interpretable Learning via Correction Vectors

论文作者

Cerrato, Mattia, Köppel, Marius, Segner, Alexander, Kramer, Stefan

论文摘要

神经网络体系结构已在公平表示学习环境中广泛使用,其目的是学习独立于敏感信息的给定向量的新表示。文献中已经提出了各种“代表性证明”技术。但是,由于神经网络本质上是不透明的,因此这些方法很难理解,这限制了它们的实用性。我们为公平表示学习提供了一个新的框架,该框架围绕“校正向量”的学习,该学习具有与给定数据向量相同的维度。然后,简单地将更正概括到原始功能,因此可以分析为对每个功能的明确罚款或奖励。我们在实验上表明,以这种方式限制的公平表示学习问题不会影响性能。

Neural network architectures have been extensively employed in the fair representation learning setting, where the objective is to learn a new representation for a given vector which is independent of sensitive information. Various "representation debiasing" techniques have been proposed in the literature. However, as neural networks are inherently opaque, these methods are hard to comprehend, which limits their usefulness. We propose a new framework for fair representation learning which is centered around the learning of "correction vectors", which have the same dimensionality as the given data vectors. The corrections are then simply summed up to the original features, and can therefore be analyzed as an explicit penalty or bonus to each feature. We show experimentally that a fair representation learning problem constrained in such a way does not impact performance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源