论文标题
COORDX:通过分裂MLP体系结构加速隐式神经表示
CoordX: Accelerating Implicit Neural Representation with a Split MLP Architecture
论文作者
论文摘要
具有多层感知器(MLP)的隐式神经表示,最近已经获得了各种各样的任务,例如新型视图合成和3D对象表示和渲染。但是,这些表示形式的一个重大挑战是,在大量输入坐标上训练和推断MLP都需要大量的计算和较长的处理时间来学习和表示图像,视频或3D对象。在这项工作中,我们旨在通过提出新的分型MLP体系结构COORDX来加速基于坐标的MLP的推理和培训,以实现隐式神经表示。使用COORDX,将初始层分开以分别学习输入坐标的每个维度。然后,中间特征由最后一层融合,以在相应的坐标点生成学习信号。这大大减少了所需的计算量,并导致训练和推理的加速较大,同时达到了与基线MLP相似的准确性。因此,该方法的目的是首先学习功能,该功能是原始信号的分解,然后融合它们生成学习的信号。我们提出的架构通常可用于许多隐式神经表示任务,而没有其他内存开销。与图像,视频和3D形状表示和渲染任务的基线模型相比,我们证明了高达2.92倍的加速度。
Implicit neural representations with multi-layer perceptrons (MLPs) have recently gained prominence for a wide variety of tasks such as novel view synthesis and 3D object representation and rendering. However, a significant challenge with these representations is that both training and inference with an MLP over a large number of input coordinates to learn and represent an image, video, or 3D object, require large amounts of computation and incur long processing times. In this work, we aim to accelerate inference and training of coordinate-based MLPs for implicit neural representations by proposing a new split MLP architecture, CoordX. With CoordX, the initial layers are split to learn each dimension of the input coordinates separately. The intermediate features are then fused by the last layers to generate the learned signal at the corresponding coordinate point. This significantly reduces the amount of computation required and leads to large speedups in training and inference, while achieving similar accuracy as the baseline MLP. This approach thus aims at first learning functions that are a decomposition of the original signal and then fusing them to generate the learned signal. Our proposed architecture can be generally used for many implicit neural representation tasks with no additional memory overheads. We demonstrate a speedup of up to 2.92x compared to the baseline model for image, video, and 3D shape representation and rendering tasks.