论文标题
通过数据中毒扩大会员资格接触
Amplifying Membership Exposure via Data Poisoning
论文作者
论文摘要
随着野外数据越来越多地参与培训阶段,机器学习应用程序变得更容易受到数据中毒攻击的影响。这种攻击通常会导致测试时间准确性降解或受控错误预测。在本文中,我们研究了第三种对数据中毒的开发 - 增加了良性训练样本的隐私泄漏风险。为此,我们演示了一组数据中毒攻击,以扩大目标类别的成员资格。我们首先提出针对监督分类算法的通用肮脏标签攻击。然后,我们在转移学习方案中提出了一种基于优化的清洁标签攻击,从而正确标记了中毒样本,并看起来“自然”以逃避人类的适应性。我们广泛评估了对计算机视觉基准测试的攻击。我们的结果表明,拟议的攻击可以大大提高成员推理精度,并以最低的总体测试时间模型性能下降。为了减轻攻击的潜在负面影响,我们还研究了可行的对策。
As in-the-wild data are increasingly involved in the training stage, machine learning applications become more susceptible to data poisoning attacks. Such attacks typically lead to test-time accuracy degradation or controlled misprediction. In this paper, we investigate the third type of exploitation of data poisoning - increasing the risks of privacy leakage of benign training samples. To this end, we demonstrate a set of data poisoning attacks to amplify the membership exposure of the targeted class. We first propose a generic dirty-label attack for supervised classification algorithms. We then propose an optimization-based clean-label attack in the transfer learning scenario, whereby the poisoning samples are correctly labeled and look "natural" to evade human moderation. We extensively evaluate our attacks on computer vision benchmarks. Our results show that the proposed attacks can substantially increase the membership inference precision with minimum overall test-time model performance degradation. To mitigate the potential negative impacts of our attacks, we also investigate feasible countermeasures.