论文标题

退火双头:深层神经网络在线校准的架构

Annealing Double-Head: An Architecture for Online Calibration of Deep Neural Networks

论文作者

Guo, Erdong, Draper, David, De Iorio, Maria

论文摘要

模型校准与模型正确预测的频率有关,不仅在统计模型设计中起着至关重要的作用,而且还具有实质性的实际应用,例如现实世界中的最佳决策。但是,已经发现,由于对预测信心的高估(或低估),现代深层神经网络通常与过度拟合密切相关。在本文中,我们提出了退火双头,这是一种简单的实施,但非常有效的架构,用于在训练过程中校准DNN。确切地说,我们构建了一个额外的校准头 - 浅的神经网络,该网络通常在正常模型中最后一个潜在层的一个潜在层顶,以将逻辑映射到对齐的置信度。此外,一种简单的退火技术,通过训练过程中的校准头动态扩展逻辑,以提高其性能。在分发和分配转变情况下,我们通过多对当代DNN架构以及愿景和语音数据集对我们的双头架构进行了详尽的评估。我们证明,我们的方法可以实现最新的模型校准性能,而无需进行后处理,同时提供了可比的预测准确性,而不是最近提出的一系列学习任务的其他提议的校准方法。

Model calibration, which is concerned with how frequently the model predicts correctly, not only plays a vital part in statistical model design, but also has substantial practical applications, such as optimal decision-making in the real world. However, it has been discovered that modern deep neural networks are generally poorly calibrated due to the overestimation (or underestimation) of predictive confidence, which is closely related to overfitting. In this paper, we propose Annealing Double-Head, a simple-to-implement but highly effective architecture for calibrating the DNN during training. To be precise, we construct an additional calibration head-a shallow neural network that typically has one latent layer-on top of the last latent layer in the normal model to map the logits to the aligned confidence. Furthermore, a simple Annealing technique that dynamically scales the logits by calibration head in training procedure is developed to improve its performance. Under both the in-distribution and distributional shift circumstances, we exhaustively evaluate our Annealing Double-Head architecture on multiple pairs of contemporary DNN architectures and vision and speech datasets. We demonstrate that our method achieves state-of-the-art model calibration performance without post-processing while simultaneously providing comparable predictive accuracy in comparison to other recently proposed calibration methods on a range of learning tasks.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源