论文标题
正规化的基于人群的培训
Regularized Evolutionary Population-Based Training
论文作者
论文摘要
深神经网络(DNN)结构和超参数的金属性已成为越来越重要的研究领域。同时,网络正则化被认为是对DNN有效培训的关键维度。但是,金属性在建立有效正则化方面的作用尚未得到充分探索。最近有证据表明,损失功能优化可以起到这种作用,但是它在计算上是完全训练的外部循环。本文提出了一种称为基于进化人群的训练(EPBT)的算法,该算法将DNN权重的训练与损失功能的金属性交织在一起。使用EPBT可以直接优化的多元泰勒扩展,它们是参数化的。这种同时适应体重和损失功能可以具有欺骗性,因此EPBT使用一种称为新颖性搏动的质量多样性启发式以及知识蒸馏来防止在训练过程中过度拟合。在CIFAR-10和SVHN图像分类基准上,EPBT会导致更快,更准确的学习。发现的超参数适应培训过程,并通过阻止过度适合标签来正规化学习任务。因此,EPBT证明了基于同时培训的正则化金属化的实例化。
Metalearning of deep neural network (DNN) architectures and hyperparameters has become an increasingly important area of research. At the same time, network regularization has been recognized as a crucial dimension to effective training of DNNs. However, the role of metalearning in establishing effective regularization has not yet been fully explored. There is recent evidence that loss-function optimization could play this role, however it is computationally impractical as an outer loop to full training. This paper presents an algorithm called Evolutionary Population-Based Training (EPBT) that interleaves the training of a DNN's weights with the metalearning of loss functions. They are parameterized using multivariate Taylor expansions that EPBT can directly optimize. Such simultaneous adaptation of weights and loss functions can be deceptive, and therefore EPBT uses a quality-diversity heuristic called Novelty Pulsation as well as knowledge distillation to prevent overfitting during training. On the CIFAR-10 and SVHN image classification benchmarks, EPBT results in faster, more accurate learning. The discovered hyperparameters adapt to the training process and serve to regularize the learning task by discouraging overfitting to the labels. EPBT thus demonstrates a practical instantiation of regularization metalearning based on simultaneous training.