论文标题
另外:按占用估算按占用率的汽车LIDAR自我选择
ALSO: Automotive Lidar Self-supervision by Occupancy estimation
论文作者
论文摘要
我们提出了一种新的自我监督方法,以预先训练在点云上运行的深度感知模型的骨干。核心思想是要以借口任务训练模型,这是对3D点进行采样的表面的重建,并使用基础潜在向量作为感知头的输入。直觉是,如果网络能够重建场景表面,只有稀疏的输入点,那么它也可能捕获了一些语义信息的片段,这些片段可用于增强实际的感知任务。该原理具有非常简单的公式,使其既易于实现,又广泛适用于大量的3D传感器和执行语义分割或对象检测的深网。实际上,它支持一条单流管线,而不是大多数对比度学习方法,从而可以对有限的资源进行培训。我们对各种自主驾驶数据集进行了广泛的实验,涉及截然不同的激光痛,以进行语义分割和对象检测。结果表明,与现有方法相比,我们的方法有效地学习有用的表示没有任何注释。代码可从https://github.com/valeoai/also获得
We propose a new self-supervised method for pre-training the backbone of deep perception models operating on point clouds. The core idea is to train the model on a pretext task which is the reconstruction of the surface on which the 3D points are sampled, and to use the underlying latent vectors as input to the perception head. The intuition is that if the network is able to reconstruct the scene surface, given only sparse input points, then it probably also captures some fragments of semantic information, that can be used to boost an actual perception task. This principle has a very simple formulation, which makes it both easy to implement and widely applicable to a large range of 3D sensors and deep networks performing semantic segmentation or object detection. In fact, it supports a single-stream pipeline, as opposed to most contrastive learning approaches, allowing training on limited resources. We conducted extensive experiments on various autonomous driving datasets, involving very different kinds of lidars, for both semantic segmentation and object detection. The results show the effectiveness of our method to learn useful representations without any annotation, compared to existing approaches. Code is available at https://github.com/valeoai/ALSO