论文标题

未靶向的后门水印:迈向无害和隐形的数据集版权保护

Untargeted Backdoor Watermark: Towards Harmless and Stealthy Dataset Copyright Protection

论文作者

Li, Yiming, Bai, Yang, Jiang, Yong, Yang, Yong, Xia, Shu-Tao, Li, Bo

论文摘要

深度神经网络(DNN)在实践中表现出了优越性。可以说,DNN的快速发展在很大程度上受益于高质量(开源)数据集,基于研究人员和开发人员可以轻松评估和改善其学习方法。由于数据收集通常是耗时甚至昂贵的,因此如何保护其版权具有重要意义,值得进一步探索。在本文中,我们重新访问数据集所有权验证。我们发现,由于仅毒药后门水印的目标性质,现有的验证方法引入了在受保护数据集中训练的DNN中引入了新的安全风险。为了减轻这一问题,在这项工作中,我们探索了不靶向的后门水印方案,在这种方案中,异常模型行为不是确定性的。具体而言,我们引入了两种分散性并证明它们的相关性,基于我们在中毒标签和清洁标签设置下设计不靶向的后门水印。我们还讨论了如何使用所提出的未靶向后门水印进行数据集所有权验证。基准数据集的实验验证了我们方法的有效性及其对现有后门防御的抵抗力。我们的代码可在\ url {https://github.com/thuyimingli/untargeted_backdoor_watermark}获得。

Deep neural networks (DNNs) have demonstrated their superiority in practice. Arguably, the rapid development of DNNs is largely benefited from high-quality (open-sourced) datasets, based on which researchers and developers can easily evaluate and improve their learning methods. Since the data collection is usually time-consuming or even expensive, how to protect their copyrights is of great significance and worth further exploration. In this paper, we revisit dataset ownership verification. We find that existing verification methods introduced new security risks in DNNs trained on the protected dataset, due to the targeted nature of poison-only backdoor watermarks. To alleviate this problem, in this work, we explore the untargeted backdoor watermarking scheme, where the abnormal model behaviors are not deterministic. Specifically, we introduce two dispersibilities and prove their correlation, based on which we design the untargeted backdoor watermark under both poisoned-label and clean-label settings. We also discuss how to use the proposed untargeted backdoor watermark for dataset ownership verification. Experiments on benchmark datasets verify the effectiveness of our methods and their resistance to existing backdoor defenses. Our codes are available at \url{https://github.com/THUYimingLi/Untargeted_Backdoor_Watermark}.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源