论文标题
自我发展的强大训练
Self-Progressing Robust Training
论文作者
论文摘要
在新的甚至对抗环境下增强模型鲁棒性是建立可信赖的机器学习系统的关键里程碑。当前的强大训练方法(例如对抗训练)明确使用“攻击”(例如$ \ ell _ {\ infty} $ - 规范界限扰动)在模型培训过程中生成对抗性示例,以改善对抗性鲁棒性。在本文中,我们采用了不同的观点,并提出了一个新的框架,称为新框架,自我创造的良好训练。在模型培训期间,发芽通过我们提出的参数标签平滑技术逐步调整培训标签分布,从而使训练无生成攻击和更具扩展性。我们还使用基于附近风险最小化的一般配方来激励发芽,其中包括许多强大的培训方法作为特殊情况。与最先进的对抗训练方法(PGD-L_INF和贸易)相比,在L_INF-NORM界攻击和各种不变性测试下,Sprout始终达到卓越的性能,并且更可扩展到大型神经网络。我们的结果为可扩展,有效和无关的鲁棒训练方法提供了新的启示。
Enhancing model robustness under new and even adversarial environments is a crucial milestone toward building trustworthy machine learning systems. Current robust training methods such as adversarial training explicitly uses an "attack" (e.g., $\ell_{\infty}$-norm bounded perturbation) to generate adversarial examples during model training for improving adversarial robustness. In this paper, we take a different perspective and propose a new framework called SPROUT, self-progressing robust training. During model training, SPROUT progressively adjusts training label distribution via our proposed parametrized label smoothing technique, making training free of attack generation and more scalable. We also motivate SPROUT using a general formulation based on vicinity risk minimization, which includes many robust training methods as special cases. Compared with state-of-the-art adversarial training methods (PGD-l_inf and TRADES) under l_inf-norm bounded attacks and various invariance tests, SPROUT consistently attains superior performance and is more scalable to large neural networks. Our results shed new light on scalable, effective and attack-independent robust training methods.