论文标题

针对联邦学习系统的数据中毒攻击

Data Poisoning Attacks Against Federated Learning Systems

论文作者

Tolpegin, Vale, Truex, Stacey, Gursoy, Mehmet Emre, Liu, Ling

论文摘要

Federated Learning(FL)是用于大规模深神经网络分布式培训的新兴范式,其中参与者的数据保留在自己的设备上,并且仅与中央服务器共享模型更新。但是,FL的分布性质引起了潜在的恶意参与者造成的新威胁。在本文中,我们研究了针对FL系统的目标数据中毒攻击,在该系统中,参与者的恶意子集旨在通过发送来自标签错误的数据的模型更新来毒化全球模型。我们首先证明,这种数据中毒攻击也会导致分类准确性大幅下降,即使有少数恶意参与者也会导致召回。我们还表明,攻击可以针对目标,即,它们只有对正在受到攻击的阶级的负面影响很大。我们还在早期/晚期训练,恶意参与者可用性的影响以及两者之间的关系中研究攻击寿命。最后,我们提出了一种防御策略,可以帮助识别FL中的恶意参与者,以规避中毒攻击,并证明其有效性。

Federated learning (FL) is an emerging paradigm for distributed training of large-scale deep neural networks in which participants' data remains on their own devices with only model updates being shared with a central server. However, the distributed nature of FL gives rise to new threats caused by potentially malicious participants. In this paper, we study targeted data poisoning attacks against FL systems in which a malicious subset of the participants aim to poison the global model by sending model updates derived from mislabeled data. We first demonstrate that such data poisoning attacks can cause substantial drops in classification accuracy and recall, even with a small percentage of malicious participants. We additionally show that the attacks can be targeted, i.e., they have a large negative impact only on classes that are under attack. We also study attack longevity in early/late round training, the impact of malicious participant availability, and the relationships between the two. Finally, we propose a defense strategy that can help identify malicious participants in FL to circumvent poisoning attacks, and demonstrate its effectiveness.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源