论文标题

反事实公平,部分已知的因果图

Counterfactual Fairness with Partially Known Causal Graph

论文作者

Zuo, Aoqi, Wei, Susan, Liu, Tongliang, Han, Bo, Zhang, Kun, Gong, Mingming

论文摘要

公平的机器学习旨在避免基于\ textit {敏感属性}(例如性别和种族)对个人或子人群的治疗。公平机器学习中的这些方法是基于因果推理确定的歧视和偏见的。尽管基于因果关系的公平学习吸引了越来越多的关注,但当前的方法假设真正的因果图是完全已知的。本文提出了一种一般方法,即当真实因果图未知时,实现反事实公平的概念。为了能够选择导致反事实公平性的功能,我们得出了条件和算法,以识别\ textIt上变量之间的祖先关系{部分定向的acyclic图(PDAG)},具体来说,是一类可导致的因果关系,可以从观测数据中学到与域知识相结合。有趣的是,我们发现可以实现反事实公平,就好像真正的因果图是完全知道的,当提供了特定的背景知识时:敏感属性在因果图中没有祖先。模拟和现实数据集的结果证明了我们方法的有效性。

Fair machine learning aims to avoid treating individuals or sub-populations unfavourably based on \textit{sensitive attributes}, such as gender and race. Those methods in fair machine learning that are built on causal inference ascertain discrimination and bias through causal effects. Though causality-based fair learning is attracting increasing attention, current methods assume the true causal graph is fully known. This paper proposes a general method to achieve the notion of counterfactual fairness when the true causal graph is unknown. To be able to select features that lead to counterfactual fairness, we derive the conditions and algorithms to identify ancestral relations between variables on a \textit{Partially Directed Acyclic Graph (PDAG)}, specifically, a class of causal DAGs that can be learned from observational data combined with domain knowledge. Interestingly, we find that counterfactual fairness can be achieved as if the true causal graph were fully known, when specific background knowledge is provided: the sensitive attributes do not have ancestors in the causal graph. Results on both simulated and real-world datasets demonstrate the effectiveness of our method.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源