Semantic segmentation approaches are typically trained on large-scale data with a closed finite set of known classes without considering unknown objects. In certain safety-critical robotics applications, especially autonomous driving, it is important to segment all objects, including those unknown at training time. We formalize the task of video class agnostic segmentation from monocular video sequences in autonomous driving to account for unknown objects. Video class agnostic segmentation can be formulated as an open-set or a motion segmentation problem. We discuss both formulations and provide datasets and benchmark different baseline approaches for both tracks. In the motion-segmentation track we benchmark real-time joint panoptic and motion instance segmentation, and evaluate the effect of ego-flow suppression. In the open-set segmentation track we evaluate baseline methods that combine appearance, and geometry to learn prototypes per semantic class. We then compare it to a model that uses an auxiliary contrastive loss to improve the discrimination between known and unknown objects. Datasets and models are publicly released at https://msiam.github.io/vca/.
翻译:在不考虑未知物体的情况下,典型的语义分割法是就大型数据进行大规模数据培训,有一套封闭的有限已知类别,不考虑未知物体。在某些安全关键的机器人应用中,特别是自主驾驶,必须分解所有物体,包括培训时未知物体。我们正式确定在自主驱动中从单视像序列中从单视像序列中进行不可知分解以核算未知物体的任务。视频类不可知分解可以作为一种开放设置或运动分解问题来拟订。我们讨论配方,并为两条轨道提供数据集和基准不同基线方法。在运动分解轨道中,我们设定实时联合光学和运动实例分解的基准,并评估自我流抑制的影响。在开放分解轨道中,我们评估将外观和地理测量相结合以学习每个语系类原型的基准方法。然后,我们将它与使用辅助对比损失来改善已知和未知对象之间区别的模式进行比较。在https://msimime.github.io/vca/上公开发布数据集和模型。