论文标题
关于主动平衡的游戏理论观点:优先的解决方案概念比NASH均衡
Game-Theoretical Perspectives on Active Equilibria: A Preferred Solution Concept over Nash Equilibria
论文作者
论文摘要
多基因学习设置本质上比单一学位学习更加困难,因为每个代理都会在共享环境中与其他同时学习代理进行交互。多种强化学习中的一种有效方法是考虑代理的学习过程,并从每个代理人的角度影响其未来的政策对理想行为。重要的是,如果每个代理通过考虑其行为对收敛策略的影响来最大程度地提高其长期奖励,则所得的多基因系统将达到主动平衡。尽管这种新的解决方案概念是一般的,因此标准解决方案概念(例如NASH平衡)是主动平衡的特殊情况,但尚不清楚主动平衡何时比其他解决方案概念是优选的平衡。在本文中,我们从游戏理论的角度分析了主动平衡,通过密切研究纳什均衡的示例。通过将主动平衡与NASH平衡直接比较这些示例,我们发现主动平衡发现比NASH平衡更有效的解决方案,得出的结论是,主动平衡是用于多种学习设置的所需解决方案。
Multiagent learning settings are inherently more difficult than single-agent learning because each agent interacts with other simultaneously learning agents in a shared environment. An effective approach in multiagent reinforcement learning is to consider the learning process of agents and influence their future policies toward desirable behaviors from each agent's perspective. Importantly, if each agent maximizes its long-term rewards by accounting for the impact of its behavior on the set of convergence policies, the resulting multiagent system reaches an active equilibrium. While this new solution concept is general such that standard solution concepts, such as a Nash equilibrium, are special cases of active equilibria, it is unclear when an active equilibrium is a preferred equilibrium over other solution concepts. In this paper, we analyze active equilibria from a game-theoretic perspective by closely studying examples where Nash equilibria are known. By directly comparing active equilibria to Nash equilibria in these examples, we find that active equilibria find more effective solutions than Nash equilibria, concluding that an active equilibrium is the desired solution for multiagent learning settings.