论文标题

自我监督的原始二元学习,以进行约束优化

Self-Supervised Primal-Dual Learning for Constrained Optimization

论文作者

Park, Seonho, Van Hentenryck, Pascal

论文摘要

本文研究了如何训练直接近似约束优化问题的最佳解决方案的机器学习模型。这是在约束下的经验风险最小化,这是具有挑战性的,因为培训必须平衡最佳和可行性条件。监督学习方法通​​常通过在大量预处理实例中训练模型来应对这一挑战。本文采用了不同的途径,并提出了原始二元学习的想法(PDL),这是一种自我监督的训练方法,不需要一组预处理的实例或用于培训和推理的优化求解器。取而代之的是,PDL模拟了增强拉格朗日方法(ALM)的轨迹,并共同训练原始和双神经网络。作为一种原始的双偶方法,PDL使用用于训练原始网络的损失函数中的约束项的实例特定惩罚。实验表明,在一组非线性优化基准上,PDL通常表现出微不足道的约束违规和较小的最佳差距,并且非常接近ALM优化。与现有方法相比,PDL在最佳差距,限制违规和培训时间方面还表现出改善或类似的性能。

This paper studies how to train machine-learning models that directly approximate the optimal solutions of constrained optimization problems. This is an empirical risk minimization under constraints, which is challenging as training must balance optimality and feasibility conditions. Supervised learning methods often approach this challenge by training the model on a large collection of pre-solved instances. This paper takes a different route and proposes the idea of Primal-Dual Learning (PDL), a self-supervised training method that does not require a set of pre-solved instances or an optimization solver for training and inference. Instead, PDL mimics the trajectory of an Augmented Lagrangian Method (ALM) and jointly trains primal and dual neural networks. Being a primal-dual method, PDL uses instance-specific penalties of the constraint terms in the loss function used to train the primal network. Experiments show that, on a set of nonlinear optimization benchmarks, PDL typically exhibits negligible constraint violations and minor optimality gaps, and is remarkably close to the ALM optimization. PDL also demonstrated improved or similar performance in terms of the optimality gaps, constraint violations, and training times compared to existing approaches.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源