论文标题
对物体检测的未靶向后门攻击
Untargeted Backdoor Attack against Object Detection
论文作者
论文摘要
最近的研究表明,在使用第三方资源(例如训练样本或骨架)培训时,深层神经网络(DNN)会面临后门威胁。该模型在预测良性样本方面具有有希望的性能,而基于对抗性的对手可以通过使用预定义的触发模式激活其后门来恶意操纵其预测。当前,大多数现有的后门攻击都是按照目标方式对图像分类进行的。在本文中,我们揭示了这些威胁在对象检测中也可能发生,从而对许多关键任务应用构成威胁风险(例如$,行人检测和智能监视系统)。具体而言,我们根据任务特征以不受限制的方式设计了一种简单而有效的唯一毒药后门攻击。我们表明,一旦通过我们的攻击将后门嵌入到目标模型中,它可能会欺骗模型失去对触发模式盖章的任何对象的检测。我们在基准数据集上进行了广泛的实验,显示了其在数字和物理世界中的有效性及其对潜在防御能力的抵抗力。
Recent studies revealed that deep neural networks (DNNs) are exposed to backdoor threats when training with third-party resources (such as training samples or backbones). The backdoored model has promising performance in predicting benign samples, whereas its predictions can be maliciously manipulated by adversaries based on activating its backdoors with pre-defined trigger patterns. Currently, most of the existing backdoor attacks were conducted on the image classification under the targeted manner. In this paper, we reveal that these threats could also happen in object detection, posing threatening risks to many mission-critical applications ($e.g.$, pedestrian detection and intelligent surveillance systems). Specifically, we design a simple yet effective poison-only backdoor attack in an untargeted manner, based on task characteristics. We show that, once the backdoor is embedded into the target model by our attack, it can trick the model to lose detection of any object stamped with our trigger patterns. We conduct extensive experiments on the benchmark dataset, showing its effectiveness in both digital and physical-world settings and its resistance to potential defenses.