Autonomous driving has achieved rapid development over the last few decades, including the machine perception as an important issue of it. Although object detection based on conventional cameras has achieved remarkable results in 2D/3D, non-visual sensors such as 3D LiDAR still have incomparable advantages in the accuracy of object position detection. However, the challenge also exists with the difficulty in properly interpreting point cloud generated by LiDAR. This paper presents a multi-modal-based online learning system for 3D LiDAR-based object classification in urban environments, including cars, cyclists and pedestrians. The proposed system aims to effectively transfer the mature detection capabilities based on visual sensors to the new model learning based on non-visual sensors through a multi-target tracker (i.e. using one sensor to train another). In particular, it integrates the Online Random Forests (ORF)~\cite{saffari2009line} method, which inherently has the abilities of fast and multi-class learning. Through experiments, we show that our system is capable of learning a high-performance model for LiDAR-based 3D object classification on-the-fly, which is especially suitable for robotics in-situ deployment while responding to the widespread challenge of insufficient detector generalization capabilities.
翻译:在过去几十年中,自主驱动取得了快速发展,包括机器认为自己是一个重要的问题。虽然基于常规摄像头的物体探测在2D/3D中取得了显著成果,但3D激光雷达等非视觉传感器在物体位置探测的准确性方面仍然具有无可比拟的优势;然而,在适当解释LIDAR产生的点云方面困难重重,也存在挑战。本文为3D LiDAR基于3DLIDAR的物体分类的城市环境,包括汽车、骑自行车者和行人,提供了一个基于3DLID的多模式在线学习系统。拟议系统的目的是通过多目标跟踪器(即使用一个传感器培训另一个传感器),将基于视觉传感器的成熟探测能力有效转让给基于非视觉传感器的新模型学习。特别是,它整合了在线随机森林(ORF)-cite{saffari2009line}方法,该方法本身具有快速和多级学习的能力。我们通过实验,表明我们的系统能够学习基于LDAR3D天物体分类的高级性模型,通过多功能在空中进行新的示范学习,这特别适合在一般机器人部署过程中进行检测。