论文标题
多权援助游戏
Multi-Principal Assistance Games
论文作者
论文摘要
援助游戏(也称为合作逆强化学习游戏)被提议作为有益AI的模型,其中机器人代理必须代表人类本金行动,但最初对人类的回报功能不确定。本文研究了多首席援助游戏,该游戏涵盖了机器人代表N人类可能有很大差异的N人行动的更一般情况。社会选择理论和投票理论中的不可能定理可以应用于此类游戏,这表明人类原则的战略行为可能会使机器人任务复杂化,以学习其回报。我们特别分析了一款强盗学徒游戏,其中人类首先采取行动来证明他们对武器的偏好,然后机器人采取行动以最大程度地提高人类回报的总和。我们探索选择次优臂的成本在多大程度上减少了误导的动机,即一种自然机制设计的形式。在这种情况下,我们提出了一种社会选择方法,该方法使用对系统的共享控制将偏好推断与社会福利优化相结合。
Assistance games (also known as cooperative inverse reinforcement learning games) have been proposed as a model for beneficial AI, wherein a robotic agent must act on behalf of a human principal but is initially uncertain about the humans payoff function. This paper studies multi-principal assistance games, which cover the more general case in which the robot acts on behalf of N humans who may have widely differing payoffs. Impossibility theorems in social choice theory and voting theory can be applied to such games, suggesting that strategic behavior by the human principals may complicate the robots task in learning their payoffs. We analyze in particular a bandit apprentice game in which the humans act first to demonstrate their individual preferences for the arms and then the robot acts to maximize the sum of human payoffs. We explore the extent to which the cost of choosing suboptimal arms reduces the incentive to mislead, a form of natural mechanism design. In this context we propose a social choice method that uses shared control of a system to combine preference inference with social welfare optimization.