论文标题

加权值得等待:贝叶斯优化具有重要的抽样

Weighting Is Worth the Wait: Bayesian Optimization with Importance Sampling

论文作者

Ariafar, Setareh, Mariet, Zelda, Elhamifar, Ehsan, Brooks, Dana, Dy, Jennifer, Snoek, Jasper

论文摘要

许多现代的机器学习模型需要对超参数进行广泛的调整才能表现良好。已经开发了多种方法,例如贝叶斯优化,以使这一过程自动化和加快这一过程。但是,调谐仍然非常昂贵,因为它通常需要反复进行全面训练模型。我们建议通过考虑每个培训示例贡献的相对信息量来加速神经网络的贝叶斯优化方法来调整神经网络的高参数。为此,我们利用重要性采样(IS);这大大提高了黑框功能评估的质量,但也必须仔细地进行运行时。将超参数搜索作为多任务贝叶斯优化问题,而不是超参数和重要性采样设计,可以实现两者的最佳状态:通过学习的参数化,即交易的评估复杂性和质量,我们可以改善贝叶斯优化的最新的运行时间和最终验证误差,并在各种数据集和复杂的神经架构中进行最终验证误差。

Many contemporary machine learning models require extensive tuning of hyperparameters to perform well. A variety of methods, such as Bayesian optimization, have been developed to automate and expedite this process. However, tuning remains extremely costly as it typically requires repeatedly fully training models. We propose to accelerate the Bayesian optimization approach to hyperparameter tuning for neural networks by taking into account the relative amount of information contributed by each training example. To do so, we leverage importance sampling (IS); this significantly increases the quality of the black-box function evaluations, but also their runtime, and so must be done carefully. Casting hyperparameter search as a multi-task Bayesian optimization problem over both hyperparameters and importance sampling design achieves the best of both worlds: by learning a parameterization of IS that trades-off evaluation complexity and quality, we improve upon Bayesian optimization state-of-the-art runtime and final validation error across a variety of datasets and complex neural architectures.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源