论文标题
自适应社会学习
Adaptive Social Learning
论文作者
论文摘要
这项工作通过引入适应的关键特征来提出一种新的社会学习策略。在社会学习中,几种分布式代理人通过以下方式不断更新他们对感兴趣现象的信念:i)直接观察他们在当地收集的流数据; ii)通过与邻居的当地合作来传播他们的信念。众所周知,传统的社会学习实施是可以很好地学习基本假设(这意味着每个人的信念在真实的假设上达到顶峰),从而在固定条件下实现了学习准确性的稳定提高。但是,这些算法在在线学习中通常遇到的非平稳条件下表现不佳,表现出重要的惯性,可以跟踪流数据中的漂移。为了解决这一差距,我们提出了一种自适应社会学习(ASL)策略,该策略依靠一个小的台阶参数来调整自适应程度。首先,我们通过稳态分析提供了学习绩效的详细表征。为了专注于小型尺寸制度,我们确定ASL策略在标准的全球可识别性假设下实现了一致的学习。我们得出了可靠的高斯近似值,以实现每个单个剂的误差概率(即选择错误的假设)。我们进行了大型偏差分析,揭示了自适应社会学习的普遍行为:误差概率随着阶梯尺寸的倒数而迅速地降低,我们表征了结果指数的学习率。其次,我们通过详细的瞬态分析来表征适应性性能,这使我们能够获得将适应时间与步进大小相关的有用的分析公式。
This work proposes a novel strategy for social learning by introducing the critical feature of adaptation. In social learning, several distributed agents update continually their belief about a phenomenon of interest through: i) direct observation of streaming data that they gather locally; and ii) diffusion of their beliefs through local cooperation with their neighbors. Traditional social learning implementations are known to learn well the underlying hypothesis (which means that the belief of every individual agent peaks at the true hypothesis), achieving steady improvement in the learning accuracy under stationary conditions. However, these algorithms do not perform well under nonstationary conditions commonly encountered in online learning, exhibiting a significant inertia to track drifts in the streaming data. In order to address this gap, we propose an Adaptive Social Learning (ASL) strategy, which relies on a small step-size parameter to tune the adaptation degree. First, we provide a detailed characterization of the learning performance by means of a steady-state analysis. Focusing on the small step-size regime, we establish that the ASL strategy achieves consistent learning under standard global identifiability assumptions. We derive reliable Gaussian approximations for the probability of error (i.e., of choosing a wrong hypothesis) at each individual agent. We carry out a large deviations analysis revealing the universal behavior of adaptive social learning: the error probabilities decrease exponentially fast with the inverse of the step-size, and we characterize the resulting exponential learning rate. Second, we characterize the adaptation performance by means of a detailed transient analysis, which allows us to obtain useful analytical formulas relating the adaptation time to the step-size.