论文标题

对抗性示例防御通过扰动分级策略

Adversarial Example Defense via Perturbation Grading Strategy

论文作者

Zhu, Shaowei, Lyu, Wanli, Li, Bin, Yin, Zhaoxia, Luo, Bin

论文摘要

深度神经网络已在许多领域广泛使用。但是,研究表明,DNN很容易受到对抗性例子的攻击,这些例子很小,这些扰动很小,并且极大地误导了DNN的正确判断。此外,即使恶意攻击者无法获得所有基础模型参数,他们也可以使用对抗性示例来攻击各种基于DNN的任务系统。研究人员提出了各种防御方法来保护DNN,例如通过添加模块进行预处理或改善模型的鲁棒性来降低对抗性实例的侵略性。但是,某些防御方法仅对小规模的例子或小扰动有效,但对具有较大扰动的对抗性实例的防御效果有限。本文通过对输入示例的扰动进行评分,将不同的防御策略分配给了不同优势的对抗性扰动。实验结果表明,所提出的方法有效地改善了防御性能。此外,所提出的方法不会修改任何任务模型,该模型可以用作预处理模块,从而大大降低了实际应用程序中的部署成本。

Deep Neural Networks have been widely used in many fields. However, studies have shown that DNNs are easily attacked by adversarial examples, which have tiny perturbations and greatly mislead the correct judgment of DNNs. Furthermore, even if malicious attackers cannot obtain all the underlying model parameters, they can use adversarial examples to attack various DNN-based task systems. Researchers have proposed various defense methods to protect DNNs, such as reducing the aggressiveness of adversarial examples by preprocessing or improving the robustness of the model by adding modules. However, some defense methods are only effective for small-scale examples or small perturbations but have limited defense effects for adversarial examples with large perturbations. This paper assigns different defense strategies to adversarial perturbations of different strengths by grading the perturbations on the input examples. Experimental results show that the proposed method effectively improves defense performance. In addition, the proposed method does not modify any task model, which can be used as a preprocessing module, which significantly reduces the deployment cost in practical applications.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源