We propose a segmentation-based bounding box generation method for omnidirectional pedestrian detection, which enables detectors to tightly fit bounding boxes to pedestrians without omnidirectional images for training. Because the appearance of pedestrians in omnidirectional images may be rotated to any angle, the performance of common pedestrian detectors is likely to be substantially degraded. Existing methods mitigate this issue by transforming images during inference or training detectors with omnidirectional images. However, the first approach substantially degrades the inference speed, and the second approach requires laborious annotations. To overcome these drawbacks, we leverage an existing large-scale dataset, whose segmentation annotations can be utilized, to generate tightly fitted bounding box annotations. We also develop a pseudo-fisheye distortion augmentation method, which further enhances the performance. Extensive analysis shows that our detector successfully fits bounding boxes to pedestrians and demonstrates substantial performance improvement.
翻译:我们建议一种基于分层的捆绑箱生成方法,用于全向行人探测,使探测器能够将盒绑在没有全向图像的行人身上进行训练。由于在全向图像中行人的外观可能旋转到任何角度,普通行人探测器的性能可能大大降低。现有的方法通过在发酵期间转换图像或用全向图像培训探测器来缓解这一问题。然而,第一种方法极大地降低了推引速度,而第二种方法则需要艰苦的注释。为了克服这些缺点,我们利用现有的大型数据集(其分解说明可以使用)来生成紧贴的捆绑框说明。我们还开发了一种假鱼眼扭曲增强功能的方法,以进一步增强性能。广泛的分析表明,我们的探测器成功地将捆绑箱对行人进行了调整,并展示了显著的性能改进。