Unmanned surface vehicles can encounter a number of varied visual circumstances during operation, some of which can be very difficult to interpret. While most cases can be solved only using color camera images, some weather and lighting conditions require additional information. To expand the available maritime data, we present a novel multimodal maritime dataset MULTIAQUA (Multimodal Aquatic Dataset). Our dataset contains synchronized, calibrated and annotated data captured by sensors of different modalities, such as RGB, thermal, IR, LIDAR, etc. The dataset is aimed at developing supervised methods that can extract useful information from these modalities in order to provide a high quality of scene interpretation regardless of potentially poor visibility conditions. To illustrate the benefits of the proposed dataset, we evaluate several multimodal methods on our difficult nighttime test set. We present training approaches that enable multimodal methods to be trained in a more robust way, thus enabling them to retain reliable performance even in near-complete darkness. Our approach allows for training a robust deep neural network only using daytime images, thus significantly simplifying data acquisition, annotation, and the training process.
翻译:无人水面舰艇在运行过程中可能遭遇多种不同的视觉环境,其中一些情况可能非常难以解析。虽然大多数情况仅使用彩色相机图像即可解决,但某些天气和光照条件需要额外信息。为扩展现有海事数据,我们提出了一个新颖的多模态海事数据集MULTIAQUA(多模态水上数据集)。该数据集包含通过RGB、热成像、红外、激光雷达等不同模态传感器采集的同步化、校准化及标注化数据。本数据集旨在开发监督式方法,以从这些模态中提取有用信息,从而在能见度可能较差的条件下仍能提供高质量的场景解析。为说明所提数据集的优势,我们在具有挑战性的夜间测试集上评估了多种多模态方法。我们提出了使多模态方法能够以更鲁棒方式进行训练的训练策略,从而使其即使在近乎完全黑暗的环境中也能保持可靠性能。我们的方法允许仅使用白天图像训练鲁棒的深度神经网络,从而显著简化了数据采集、标注及训练过程。