论文标题
带有互补标签的对抗性培训:逐渐有用的攻击受益
Adversarial Training with Complementary Labels: On the Benefit of Gradually Informative Attacks
论文作者
论文摘要
具有不完美监督的对抗训练(AT)是显着的,但受到有限的关注。为了朝着更实际的方案迈进,我们探索了一个全新但具有挑战性的环境,即带有互补标签(CLS),该设置指定了数据样本不属于的类。但是,AT与CLS现有方法的直接组合会导致失败,但并不能在两个阶段训练的简单基线上进行。在本文中,我们进一步探讨了现象,并确定CLS的基本挑战是棘手的对抗优化和低质量的对抗性例子。为了解决上述问题,我们提出了一种新的学习策略,使用逐渐信息攻击,其中包括两个关键组成部分:1)热身攻击(热身)轻轻提高对抗性扰动预算,以缓解使用CLS的对抗性优化; 2)伪标签攻击(PLA)将逐步信息的模型预测纳入校正的互补损失中。进行了广泛的实验,以证明我们方法对一系列基准数据集的有效性。该代码可公开可用:https://github.com/royalskye/atcl。
Adversarial training (AT) with imperfect supervision is significant but receives limited attention. To push AT towards more practical scenarios, we explore a brand new yet challenging setting, i.e., AT with complementary labels (CLs), which specify a class that a data sample does not belong to. However, the direct combination of AT with existing methods for CLs results in consistent failure, but not on a simple baseline of two-stage training. In this paper, we further explore the phenomenon and identify the underlying challenges of AT with CLs as intractable adversarial optimization and low-quality adversarial examples. To address the above problems, we propose a new learning strategy using gradually informative attacks, which consists of two critical components: 1) Warm-up Attack (Warm-up) gently raises the adversarial perturbation budgets to ease the adversarial optimization with CLs; 2) Pseudo-Label Attack (PLA) incorporates the progressively informative model predictions into a corrected complementary loss. Extensive experiments are conducted to demonstrate the effectiveness of our method on a range of benchmarked datasets. The code is publicly available at: https://github.com/RoyalSkye/ATCL.