论文标题

混乱是一个阶梯:通过增强重叠对对比学习的新理论理解

Chaos is a Ladder: A New Theoretical Understanding of Contrastive Learning via Augmentation Overlap

论文作者

Wang, Yifei, Zhang, Qi, Wang, Yisen, Yang, Jiansheng, Lin, Zhouchen

论文摘要

最近,对比学习已成为大规模自学学习的一种有希望的方法。但是,对其运作方式的理论理解仍然不清楚。在本文中,我们提出了有关下游性能的新保证,而不诉诸于以前的工作中广泛采用的有条件独立性假设,但实际上几乎不存在。我们的新理论取决于洞察力,即在积极的数据增强下,不同类样品的支持将变得更加重叠,因此,简单地对齐正样本(同一样本的增强视图)可以使对比性学习群集在阶层内样本在一起。从理论上讲,基于这种增强重叠的视角,我们在较弱的假设下获得了下游性能的渐近封闭界限,并且从经验上讲,我们提出了一个无监督的模型选择度量弧,与下游准确性很好地符合。我们的理论提出了对对比学习的另一种理解:对齐正样本的作用更像是代孕任务,而不是一个最终目标,而重叠的增强观点(即混乱)为对比学习创造了一个阶梯,以逐步学习课堂分离的代表。用于计算弧的代码可在https://github.com/zhangq327/arc上获得。

Recently, contrastive learning has risen to be a promising approach for large-scale self-supervised learning. However, theoretical understanding of how it works is still unclear. In this paper, we propose a new guarantee on the downstream performance without resorting to the conditional independence assumption that is widely adopted in previous work but hardly holds in practice. Our new theory hinges on the insight that the support of different intra-class samples will become more overlapped under aggressive data augmentations, thus simply aligning the positive samples (augmented views of the same sample) could make contrastive learning cluster intra-class samples together. Based on this augmentation overlap perspective, theoretically, we obtain asymptotically closed bounds for downstream performance under weaker assumptions, and empirically, we propose an unsupervised model selection metric ARC that aligns well with downstream accuracy. Our theory suggests an alternative understanding of contrastive learning: the role of aligning positive samples is more like a surrogate task than an ultimate goal, and the overlapped augmented views (i.e., the chaos) create a ladder for contrastive learning to gradually learn class-separated representations. The code for computing ARC is available at https://github.com/zhangq327/ARC.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源