论文标题

变压器的端到端符号回归

End-to-end symbolic regression with transformers

论文作者

Kamienny, Pierre-Alexandre, d'Ascoli, Stéphane, Lample, Guillaume, Charton, François

论文摘要

符号回归是从观察值的观察中预测函数的数学表达的任务,是一个困难的任务,通常涉及两步过程:预测表达式的“骨架”,直到选择数值常数,然后通过通过非convex损耗函数来优化常数。主要的方法是遗传编程,它通过迭代该子例程大量次数来演变候选者。神经网络最近被任命在一次尝试中预测正确的骨骼,但功能较低。在本文中,我们挑战了这两个步骤的过程,并任命“变压器”直接预测包括常数的完整数学表达。随后,可以通过将预测的常数作为知情初始化来完善预测的常数。我们提供消融,以表明这种端到端的方法会产生更好的结果,有时即使没有完善步骤也是如此。我们对SRBENCH基准测试的问题评估了模型,并表明我们的模型以更快的推断速度来处理最先进的基因编程的性能。

Symbolic regression, the task of predicting the mathematical expression of a function from the observation of its values, is a difficult task which usually involves a two-step procedure: predicting the "skeleton" of the expression up to the choice of numerical constants, then fitting the constants by optimizing a non-convex loss function. The dominant approach is genetic programming, which evolves candidates by iterating this subroutine a large number of times. Neural networks have recently been tasked to predict the correct skeleton in a single try, but remain much less powerful. In this paper, we challenge this two-step procedure, and task a Transformer to directly predict the full mathematical expression, constants included. One can subsequently refine the predicted constants by feeding them to the non-convex optimizer as an informed initialization. We present ablations to show that this end-to-end approach yields better results, sometimes even without the refinement step. We evaluate our model on problems from the SRBench benchmark and show that our model approaches the performance of state-of-the-art genetic programming with several orders of magnitude faster inference.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源