Multi-beam LiDAR sensors, as used on autonomous vehicles and mobile robots, acquire sequences of 3D range scans ("frames"). Each frame covers the scene sparsely, due to limited angular scanning resolution and occlusion. The sparsity restricts the performance of downstream processes like semantic segmentation or surface reconstruction. Luckily, when the sensor moves, frames are captured from a sequence of different viewpoints. This provides complementary information and, when accumulated in a common scene coordinate frame, yields a denser sampling and a more complete coverage of the underlying 3D scene. However, often the scanned scenes contain moving objects. Points on those objects are not correctly aligned by just undoing the scanner's ego-motion. In the present paper, we explore multi-frame point cloud accumulation as a mid-level representation of 3D scan sequences, and develop a method that exploits inductive biases of outdoor street scenes, including their geometric layout and object-level rigidity. Compared to state-of-the-art scene flow estimators, our proposed approach aims to align all 3D points in a common reference frame correctly accumulating the points on the individual objects. Our approach greatly reduces the alignment errors on several benchmark datasets. Moreover, the accumulated point clouds benefit high-level tasks like surface reconstruction.
翻译:自动飞行器和移动机器人上使用的多波束激光雷达传感器获得3D范围扫描(“框架”)的序列。每个框架覆盖的场景很少。由于角扫描分辨率和封闭性有限,每个框架覆盖的场景很少。宽度限制了下游过程的性能,例如语义分割或表面重建。幸运的是,当传感器移动时,框架从不同观点的序列中采集。这提供了补充信息,当在共同的场景协调框架内积累时,可以产生更密集的取样,并更完整地覆盖基底的3D场景。然而,扫描场景往往包含移动的物体。这些物体上的点没有通过仅仅解开扫描仪的自我感动来正确对齐。在本文件中,我们探索多框架点云积聚,作为3D扫描序列的中层表示,并开发一种方法,利用户外街头场景的感动偏向,包括几何形状的布局和对象的僵硬度。与最先进的场景图示显示器相比,我们拟议的方法旨在将所有3D点与所有点相对齐,只是通过解扫描仪的自我触控。在共同的高度参照点上,我们的拟议方法减少了一些地面调整了地平标点。