论文标题

并行矩形翻转攻击:基于查询的黑框攻击对象检测

Parallel Rectangle Flip Attack: A Query-based Black-box Attack against Object Detection

论文作者

Liang, Siyuan, Wu, Baoyuan, Fan, Yanbo, Wei, Xingxing, Cao, Xiaochun

论文摘要

对象检测已被广泛用于许多安全至关重要的任务,例如自动驾驶。但是,其对对抗性示例的脆弱性尚未得到充分研究,尤其是在黑盒攻击的实际情况下,攻击者只能访问攻击模型返回的预测边界盒和TOP-1分数的查询反馈。与黑框攻击到图像分类相比,黑框攻击要检测到两个主要挑战。首先,即使成功攻击了一个边界框,也可以在攻击的边界框附近检测到另一个亚最佳边界框。其次,有多个边界盒,导致攻击成本很高。为了应对这些挑战,我们通过随机搜索提出了平行的矩形翻转攻击(PRFA)。我们解释了我们的方法与图〜\ ref {fig1}中的其他攻击之间的区别。具体而言,我们在每个矩形贴片中产生扰动,以避免在受攻击区域附近的次优检测。此外,通过观察到,对抗扰动主要是在白色盒子攻击下围绕对象的轮廓和关键点定位,攻击矩形的搜索空间将缩小以提高攻击效率。此外,我们开发了一种并行的机制,即同时攻击多个矩形以进一步加速攻击过程。广泛的实验表明,我们的方法可以有效,有效地攻击各种流行的对象探测器,包括基于锚和锚定的,并生成可转移的对抗性示例。

Object detection has been widely used in many safety-critical tasks, such as autonomous driving. However, its vulnerability to adversarial examples has not been sufficiently studied, especially under the practical scenario of black-box attacks, where the attacker can only access the query feedback of predicted bounding-boxes and top-1 scores returned by the attacked model. Compared with black-box attack to image classification, there are two main challenges in black-box attack to detection. Firstly, even if one bounding-box is successfully attacked, another sub-optimal bounding-box may be detected near the attacked bounding-box. Secondly, there are multiple bounding-boxes, leading to very high attack cost. To address these challenges, we propose a Parallel Rectangle Flip Attack (PRFA) via random search. We explain the difference between our method with other attacks in Fig.~\ref{fig1}. Specifically, we generate perturbations in each rectangle patch to avoid sub-optimal detection near the attacked region. Besides, utilizing the observation that adversarial perturbations mainly locate around objects' contours and critical points under white-box attacks, the search space of attacked rectangles is reduced to improve the attack efficiency. Moreover, we develop a parallel mechanism of attacking multiple rectangles simultaneously to further accelerate the attack process. Extensive experiments demonstrate that our method can effectively and efficiently attack various popular object detectors, including anchor-based and anchor-free, and generate transferable adversarial examples.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源