论文标题

神经回报机:预测团队成员之间的公平和稳定的收益分配

Neural Payoff Machines: Predicting Fair and Stable Payoff Allocations Among Team Members

论文作者

Cornelisse, Daphne, Rood, Thomas, Malinowski, Mateusz, Bachrach, Yoram, Kachman, Tal

论文摘要

在许多多代理设置中,参与者可以组建团队以实现可能超过个人能力的集体成果。衡量代理商的相对贡献并分配促进长期合作的奖励份额是艰巨的任务。合作游戏理论提供了识别分配方案(例如莎普利价值)的解决方案概念,这些概念公平地反映了个人对团队或核心表现的贡献,从而减少了代理人放弃团队的动机。此类方法的应用包括确定有影响力的特征并分享合资企业或团队成立的成本。不幸的是,即使在受限设置中,使用这些解决方案也需要解决计算障碍,因为它们很难计算。在这项工作中,我们展示了如何通过训练神经网络提出公平稳定的回报分配来将合作游戏理论解决方案蒸馏成一个学识渊博的模型。我们表明,我们的方法创建的模型可以推广到远离训练分布的游戏,并且可以预测比训练期间观察到的更多玩家的解决方案。我们框架的一个重要应用是可以解释的AI:我们的方法可用于加快在许多情况下的Shapley价值计算。

In many multi-agent settings, participants can form teams to achieve collective outcomes that may far surpass their individual capabilities. Measuring the relative contributions of agents and allocating them shares of the reward that promote long-lasting cooperation are difficult tasks. Cooperative game theory offers solution concepts identifying distribution schemes, such as the Shapley value, that fairly reflect the contribution of individuals to the performance of the team or the Core, which reduces the incentive of agents to abandon their team. Applications of such methods include identifying influential features and sharing the costs of joint ventures or team formation. Unfortunately, using these solutions requires tackling a computational barrier as they are hard to compute, even in restricted settings. In this work, we show how cooperative game-theoretic solutions can be distilled into a learned model by training neural networks to propose fair and stable payoff allocations. We show that our approach creates models that can generalize to games far from the training distribution and can predict solutions for more players than observed during training. An important application of our framework is Explainable AI: our approach can be used to speed-up Shapley value computations on many instances.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源