As technology advances in autonomous mobile robots, mobile service robots have been actively used more and more for various purposes. Especially, serving robots have been not surprising products anymore since the COVID-19 pandemic. One of the practical problems in operating serving a robot is that it often fails to estimate its pose on a map that it moves around. Whenever the failure happens, servers should bring the serving robot to its initial location and reboot it manually. In this paper, we focus on end-to-end relocalization of serving robots to address the problem. It is to predict robot pose directly from only the onboard sensor data using neural networks. In particular, we propose a deep neural network architecture for the relocalization based on camera-2D LiDAR sensor fusion. We call the proposed method FusionLoc. In the proposed method, the multi-head self-attention complements different types of information captured by the two sensors to regress the robot pose. Our experiments on a dataset collected by a commercial serving robot demonstrate that FusionLoc can provide better performances than previous end-to-end relocalization methods taking only a single image or a 2D LiDAR point cloud as well as a straightforward fusion method concatenating their features.
翻译:随着自动移动机器人技术的不断发展,移动服务机器人在各种用途中得到了越来越广泛的应用。特别是在COVID-19大流行期间,服务机器人已经成为人们日常生活中的常见产品。服务机器人运行中的实际问题之一是它经常无法估计它在地图上移动的位置。每当发生故障时,服务器都需要将服务机器人带回初始位置并手动重新启动它。本文专注于针对该问题的服务机器人的端到端重定位。它是直接从仅通过神经网络使用机载传感器数据来预测机器人姿态的技术。特别地,我们提出了一种基于相机-2D LiDAR传感器融合的重定位深度神经网络架构,我们称之为FusionLoc。在所提出的方法中,多头自注意力为两个不同传感器捕获的不同类型的信息提供补充,以回归机器人姿态。我们在商业服务机器人收集的数据集上进行的实验证明,与仅使用单个图像或2D LiDAR点云的先前端到端重定位方法以及简单融合其特征的方法相比, FusionLoc可以提供更好的性能。