论文标题
3D场景的本地隐式网格表示
Local Implicit Grid Representations for 3D Scenes
论文作者
论文摘要
从数据中学到的形状先验通常用于从部分或嘈杂数据中重建3D对象。然而,由于典型的3D自动编码器无法处理其规模,复杂性或多样性,因此没有这样的形状先验可用于室内场景。在本文中,我们介绍了本地隐式网格表示,这是一种新的3D形状表示形式,旨在可扩展性和通用性。激励人心的想法是,大多数3D表面都以某种规模共享几何细节 - 即,比整个对象小,比小斑块大。我们训练一个自动编码器,以学习该尺寸的3D形状的当地农作物的嵌入。然后,我们将解码器用作形状优化中的组件,该组件在重叠作物的常规网格上求解一组潜在代码,以使解码的局部形状的插值与部分或嘈杂的观测值匹配。我们证明了从稀疏点观测结果中提出的3D表面重建方法的值,与替代方法相比,结果明显更好。
Shape priors learned from data are commonly used to reconstruct 3D objects from partial or noisy data. Yet no such shape priors are available for indoor scenes, since typical 3D autoencoders cannot handle their scale, complexity, or diversity. In this paper, we introduce Local Implicit Grid Representations, a new 3D shape representation designed for scalability and generality. The motivating idea is that most 3D surfaces share geometric details at some scale -- i.e., at a scale smaller than an entire object and larger than a small patch. We train an autoencoder to learn an embedding of local crops of 3D shapes at that size. Then, we use the decoder as a component in a shape optimization that solves for a set of latent codes on a regular grid of overlapping crops such that an interpolation of the decoded local shapes matches a partial or noisy observation. We demonstrate the value of this proposed approach for 3D surface reconstruction from sparse point observations, showing significantly better results than alternative approaches.