论文标题

通过不同的RNN吸引者实施不同导航任务的归纳偏见

Implementing Inductive bias for different navigation tasks through diverse RNN attractors

论文作者

Xu, Tie, Barak, Omri

论文摘要

导航对于动物行为至关重要,被认为需要对外部环境的内部表示,称为认知图。这种表示的精确形式通常被认为是空间的度量表示。但是,内部表示是根据其对给定任务的绩效的贡献来判断的,因此在不同类型的导航任务之间可能会有所不同。在这里,我们训练一个经常性的神经网络,该神经网络控制一个在简单环境中执行多个导航任务的代理。为了关注内部表示形式,我们将学习分为任务不合时宜的预训练阶段,该阶段修改了内部连接性和控制网络输出的特定任务Q学习阶段。我们表明,预训练塑造了网络的吸引子景观,从而导致连续的吸引子,离散的吸引子或无序状态。这些结构会在Q学习阶段引起偏见,从而导致与公制和拓扑规律相对应的任务的性能模式。通过将两种类型的网络组合在模块化结构中,我们可以为两个规律性获得更好的性能。我们的结果表明,在经常性网络中,归纳偏差采用吸引子景观的形式 - 可以通过预训练来塑造并使用动力学系统方法对其进行分析。此外,我们证明了非金属表示对导航任务很有用,并且它们与度量表示的组合会导致挠性多任务学习。

Navigation is crucial for animal behavior and is assumed to require an internal representation of the external environment, termed a cognitive map. The precise form of this representation is often considered to be a metric representation of space. An internal representation, however, is judged by its contribution to performance on a given task, and may thus vary between different types of navigation tasks. Here we train a recurrent neural network that controls an agent performing several navigation tasks in a simple environment. To focus on internal representations, we split learning into a task-agnostic pre-training stage that modifies internal connectivity and a task-specific Q learning stage that controls the network's output. We show that pre-training shapes the attractor landscape of the networks, leading to either a continuous attractor, discrete attractors or a disordered state. These structures induce bias onto the Q-Learning phase, leading to a performance pattern across the tasks corresponding to metric and topological regularities. By combining two types of networks in a modular structure, we could get better performance for both regularities. Our results show that, in recurrent networks, inductive bias takes the form of attractor landscapes -- which can be shaped by pre-training and analyzed using dynamical systems methods. Furthermore, we demonstrate that non-metric representations are useful for navigation tasks, and their combination with metric representation leads to flexibile multiple-task learning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源