论文标题
如何在联合学习中结合变分的贝叶斯网络
How to Combine Variational Bayesian Networks in Federated Learning
论文作者
论文摘要
联合学习使多个数据中心能够协作培训中心模型,而无需公开任何机密数据。即使确定性模型能够执行高预测准确性,但它们缺乏校准和量化不确定性的能力对于安全至关重要的应用是有问题的。与确定性模型不同的是,诸如贝叶斯神经网络之类的概率模型相对良好,能够量化不确定性以及其竞争性预测准确性。两种方法都出现在联邦学习框架中。但是,确定性模型的聚合方案不能直接应用于概率模型,因为权重与分布而不是点估计相对应。在这项工作中,我们研究了各种聚合方案对变异贝叶斯神经网络的影响。通过三个图像分类数据集的经验结果,我们观察到汇总分布的扩散程度是学习过程中的重要因素。因此,我们提出了有关如何在联合学习中结合变异贝叶斯网络的问题的调查,同时为不同的聚合设置提供了基准。
Federated Learning enables multiple data centers to train a central model collaboratively without exposing any confidential data. Even though deterministic models are capable of performing high prediction accuracy, their lack of calibration and capability to quantify uncertainty is problematic for safety-critical applications. Different from deterministic models, probabilistic models such as Bayesian neural networks are relatively well-calibrated and able to quantify uncertainty alongside their competitive prediction accuracy. Both of the approaches appear in the federated learning framework; however, the aggregation scheme of deterministic models cannot be directly applied to probabilistic models since weights correspond to distributions instead of point estimates. In this work, we study the effects of various aggregation schemes for variational Bayesian neural networks. With empirical results on three image classification datasets, we observe that the degree of spread for an aggregated distribution is a significant factor in the learning process. Hence, we present an investigation on the question of how to combine variational Bayesian networks in federated learning, while providing benchmarks for different aggregation settings.