论文标题
合作人工智能
Cooperative Artificial Intelligence
论文作者
论文摘要
将来,人工学习的代理人可能在我们的社会中变得越来越普遍。他们将在包括社会困境在内的各种复杂环境中与其他学习者和人类互动。我们认为,需要研究游戏理论与人工智能之间的交集,其目的是实现可以很好地应对社会困境的合作人工智能。我们考虑了外部代理如何通过观察学习者的行动来分发额外的奖励和惩罚来促进人工学习者之间的合作的问题。我们提出了一个规则,用于自动学习如何通过考虑每个代理的预期参数更新来创建正确的激励措施。使用此学习规则会导致与高社会福利在矩阵游戏中的合作,否则代理商将学会以很高的可能性学习缺陷。我们表明,即使在给定数量的情节之后关闭计划代理人,在某些游戏中,由此产生的合作结果稳定,而其他游戏则需要持续干预才能保持相互合作。最后,我们反思了多代理强化学习的目标首先是什么,并讨论了建立合作AI目标的必要构件。
In the future, artificial learning agents are likely to become increasingly widespread in our society. They will interact with both other learning agents and humans in a variety of complex settings including social dilemmas. We argue that there is a need for research on the intersection between game theory and artificial intelligence, with the goal of achieving cooperative artificial intelligence that can navigate social dilemmas well. We consider the problem of how an external agent can promote cooperation between artificial learners by distributing additional rewards and punishments based on observing the actions of the learners. We propose a rule for automatically learning how to create the right incentives by considering the anticipated parameter updates of each agent. Using this learning rule leads to cooperation with high social welfare in matrix games in which the agents would otherwise learn to defect with high probability. We show that the resulting cooperative outcome is stable in certain games even if the planning agent is turned off after a given number of episodes, while other games require ongoing intervention to maintain mutual cooperation. Finally, we reflect on what the goals of multi-agent reinforcement learning should be in the first place, and discuss the necessary building blocks towards the goal of building cooperative AI.