论文标题
Imagenet上的鲁棒性是否转移到下游任务?
Does Robustness on ImageNet Transfer to Downstream Tasks?
论文作者
论文摘要
随着干净的成像网精度接近其上限,研究界越来越关注分配变化的稳健精度。尽管已经提出了多种方法来鲁棒性神经网络,但这些技术通常针对接受成像网分类训练的模型。同时,使用ImageNet预估计的骨干进行下游任务是一种常见的做法,例如来自不同域的对象检测,语义分割和图像分类。这提出了一个问题:这些强大的图像分类器可以将鲁棒性转移到下游任务吗?为了进行对象检测和语义细分,我们发现一种香草丝变形金刚,这是一种量身定制的视觉变压器,适合密集的预测任务,比卷积神经网络更好地转移鲁棒性,该卷积神经网络被训练为损坏的Imagenet损坏版本。对于CIFAR10分类,我们发现在完全微调时,具有鲁棒化的模型不会保留鲁棒性。这些发现表明,当前的鲁棒化技术倾向于强调成像网评估。此外,当我们考虑转移学习时,网络体系结构是强大的鲁棒性来源。
As clean ImageNet accuracy nears its ceiling, the research community is increasingly more concerned about robust accuracy under distributional shifts. While a variety of methods have been proposed to robustify neural networks, these techniques often target models trained on ImageNet classification. At the same time, it is a common practice to use ImageNet pretrained backbones for downstream tasks such as object detection, semantic segmentation, and image classification from different domains. This raises a question: Can these robust image classifiers transfer robustness to downstream tasks? For object detection and semantic segmentation, we find that a vanilla Swin Transformer, a variant of Vision Transformer tailored for dense prediction tasks, transfers robustness better than Convolutional Neural Networks that are trained to be robust to the corrupted version of ImageNet. For CIFAR10 classification, we find that models that are robustified for ImageNet do not retain robustness when fully fine-tuned. These findings suggest that current robustification techniques tend to emphasize ImageNet evaluations. Moreover, network architecture is a strong source of robustness when we consider transfer learning.